var/home/core/zuul-output/0000755000175000017500000000000015071477201014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071514754015502 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005440514015071514745017711 0ustar rootrootOct 08 15:13:09 crc systemd[1]: Starting Kubernetes Kubelet... Oct 08 15:13:09 crc restorecon[4700]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:09 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 15:13:10 crc restorecon[4700]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 08 15:13:11 crc kubenswrapper[4945]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 15:13:11 crc kubenswrapper[4945]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 08 15:13:11 crc kubenswrapper[4945]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 15:13:11 crc kubenswrapper[4945]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 15:13:11 crc kubenswrapper[4945]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 08 15:13:11 crc kubenswrapper[4945]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.484080 4945 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491921 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491967 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491972 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491977 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491983 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491989 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491993 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.491998 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492002 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492008 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492014 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492021 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492027 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492033 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492039 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492044 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492048 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492053 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492062 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492068 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492074 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492079 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492084 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492089 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492093 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492097 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492102 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492123 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492148 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492154 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492161 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492166 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492172 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492177 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492183 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492188 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492192 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492198 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492203 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492208 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492213 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492217 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492221 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492226 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492230 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492236 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492240 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492246 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492251 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492258 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492264 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492269 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492274 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492278 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492283 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492288 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492292 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492297 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492301 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492306 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492310 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492315 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492320 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492324 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492329 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492334 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492338 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492342 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492346 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492351 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.492355 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492461 4945 flags.go:64] FLAG: --address="0.0.0.0" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492476 4945 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492485 4945 flags.go:64] FLAG: --anonymous-auth="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492493 4945 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492502 4945 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492507 4945 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492517 4945 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492525 4945 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492530 4945 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492535 4945 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492540 4945 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492549 4945 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492555 4945 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492560 4945 flags.go:64] FLAG: --cgroup-root="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492565 4945 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492570 4945 flags.go:64] FLAG: --client-ca-file="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492575 4945 flags.go:64] FLAG: --cloud-config="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492580 4945 flags.go:64] FLAG: --cloud-provider="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492584 4945 flags.go:64] FLAG: --cluster-dns="[]" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492592 4945 flags.go:64] FLAG: --cluster-domain="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492597 4945 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492602 4945 flags.go:64] FLAG: --config-dir="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492607 4945 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492612 4945 flags.go:64] FLAG: --container-log-max-files="5" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492620 4945 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492626 4945 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492632 4945 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492638 4945 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492643 4945 flags.go:64] FLAG: --contention-profiling="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492649 4945 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492654 4945 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492659 4945 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492664 4945 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492671 4945 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492678 4945 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492683 4945 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492688 4945 flags.go:64] FLAG: --enable-load-reader="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492694 4945 flags.go:64] FLAG: --enable-server="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492700 4945 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492707 4945 flags.go:64] FLAG: --event-burst="100" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492712 4945 flags.go:64] FLAG: --event-qps="50" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492717 4945 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492723 4945 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492729 4945 flags.go:64] FLAG: --eviction-hard="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492743 4945 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492748 4945 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492753 4945 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492760 4945 flags.go:64] FLAG: --eviction-soft="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492765 4945 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492771 4945 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492776 4945 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492781 4945 flags.go:64] FLAG: --experimental-mounter-path="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492787 4945 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492792 4945 flags.go:64] FLAG: --fail-swap-on="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492797 4945 flags.go:64] FLAG: --feature-gates="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492804 4945 flags.go:64] FLAG: --file-check-frequency="20s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492809 4945 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492815 4945 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492821 4945 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492826 4945 flags.go:64] FLAG: --healthz-port="10248" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492831 4945 flags.go:64] FLAG: --help="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492837 4945 flags.go:64] FLAG: --hostname-override="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492842 4945 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492849 4945 flags.go:64] FLAG: --http-check-frequency="20s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492855 4945 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492861 4945 flags.go:64] FLAG: --image-credential-provider-config="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492868 4945 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492875 4945 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492881 4945 flags.go:64] FLAG: --image-service-endpoint="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492887 4945 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492893 4945 flags.go:64] FLAG: --kube-api-burst="100" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492898 4945 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492904 4945 flags.go:64] FLAG: --kube-api-qps="50" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492909 4945 flags.go:64] FLAG: --kube-reserved="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492914 4945 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492919 4945 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492925 4945 flags.go:64] FLAG: --kubelet-cgroups="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492929 4945 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492945 4945 flags.go:64] FLAG: --lock-file="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492950 4945 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492955 4945 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492961 4945 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492969 4945 flags.go:64] FLAG: --log-json-split-stream="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492974 4945 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492979 4945 flags.go:64] FLAG: --log-text-split-stream="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492984 4945 flags.go:64] FLAG: --logging-format="text" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492989 4945 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.492995 4945 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493208 4945 flags.go:64] FLAG: --manifest-url="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493214 4945 flags.go:64] FLAG: --manifest-url-header="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493224 4945 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493229 4945 flags.go:64] FLAG: --max-open-files="1000000" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493237 4945 flags.go:64] FLAG: --max-pods="110" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493242 4945 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493248 4945 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493253 4945 flags.go:64] FLAG: --memory-manager-policy="None" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493258 4945 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493263 4945 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493269 4945 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493274 4945 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493290 4945 flags.go:64] FLAG: --node-status-max-images="50" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493296 4945 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493302 4945 flags.go:64] FLAG: --oom-score-adj="-999" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493307 4945 flags.go:64] FLAG: --pod-cidr="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493311 4945 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493321 4945 flags.go:64] FLAG: --pod-manifest-path="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493326 4945 flags.go:64] FLAG: --pod-max-pids="-1" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493331 4945 flags.go:64] FLAG: --pods-per-core="0" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493337 4945 flags.go:64] FLAG: --port="10250" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493342 4945 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493350 4945 flags.go:64] FLAG: --provider-id="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493355 4945 flags.go:64] FLAG: --qos-reserved="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493360 4945 flags.go:64] FLAG: --read-only-port="10255" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493365 4945 flags.go:64] FLAG: --register-node="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493371 4945 flags.go:64] FLAG: --register-schedulable="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493376 4945 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493387 4945 flags.go:64] FLAG: --registry-burst="10" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493392 4945 flags.go:64] FLAG: --registry-qps="5" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493397 4945 flags.go:64] FLAG: --reserved-cpus="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493403 4945 flags.go:64] FLAG: --reserved-memory="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493411 4945 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493416 4945 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493421 4945 flags.go:64] FLAG: --rotate-certificates="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493427 4945 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493432 4945 flags.go:64] FLAG: --runonce="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493438 4945 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493443 4945 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493449 4945 flags.go:64] FLAG: --seccomp-default="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493454 4945 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493460 4945 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493466 4945 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493471 4945 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493477 4945 flags.go:64] FLAG: --storage-driver-password="root" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493482 4945 flags.go:64] FLAG: --storage-driver-secure="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493488 4945 flags.go:64] FLAG: --storage-driver-table="stats" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493495 4945 flags.go:64] FLAG: --storage-driver-user="root" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493501 4945 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493508 4945 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493514 4945 flags.go:64] FLAG: --system-cgroups="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493519 4945 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493530 4945 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493536 4945 flags.go:64] FLAG: --tls-cert-file="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493544 4945 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493556 4945 flags.go:64] FLAG: --tls-min-version="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493562 4945 flags.go:64] FLAG: --tls-private-key-file="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493567 4945 flags.go:64] FLAG: --topology-manager-policy="none" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493573 4945 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493580 4945 flags.go:64] FLAG: --topology-manager-scope="container" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493588 4945 flags.go:64] FLAG: --v="2" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493596 4945 flags.go:64] FLAG: --version="false" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493605 4945 flags.go:64] FLAG: --vmodule="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493614 4945 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.493621 4945 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493787 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493795 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493802 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493807 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493813 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493817 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493824 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493830 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493835 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493841 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493847 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493853 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493859 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493865 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493871 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493876 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493883 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493888 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493892 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493897 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493902 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493909 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493914 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493918 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493923 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493928 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493932 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493936 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493942 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493948 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493953 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493958 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493964 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493969 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493974 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493979 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493983 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493988 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493993 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.493998 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494004 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494009 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494014 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494019 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494024 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494028 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494034 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494041 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494047 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494052 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494057 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494063 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494068 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494077 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494083 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494089 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494095 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494100 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494121 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494127 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494131 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494137 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494143 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494148 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494153 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494158 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494163 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494167 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494173 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494177 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.494182 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.494198 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.510100 4945 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.510184 4945 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510325 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510339 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510350 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510359 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510368 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510376 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510385 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510397 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510409 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510418 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510426 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510435 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510445 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510454 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510462 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510470 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510518 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510528 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510537 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510546 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510554 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510562 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510570 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510578 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510586 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510594 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510602 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510609 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510618 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510625 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510633 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510642 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510653 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510668 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510686 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510698 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510707 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510716 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510727 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510736 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510744 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510753 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510764 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510775 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510783 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510791 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510802 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510814 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510823 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510831 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510839 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510847 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510854 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510862 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510870 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510878 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510886 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510894 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510902 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510910 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510917 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510925 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510933 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510941 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510949 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510959 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510967 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510975 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510983 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.510991 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511000 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.511013 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511271 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511286 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511296 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511306 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511315 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511324 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511333 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511342 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511350 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511360 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511368 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511376 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511384 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511392 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511400 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511408 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511416 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511424 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511433 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511440 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511451 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511463 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511473 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511484 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511493 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511502 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511510 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511519 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511528 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511536 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511544 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511552 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511560 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511568 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511577 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511585 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511593 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511600 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511608 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511616 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511624 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511634 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511647 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511659 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511673 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511684 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511695 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511703 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511712 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511721 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511729 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511737 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511745 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511753 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511762 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511770 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511777 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511785 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511795 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511804 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511813 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511821 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511830 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511838 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511846 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511856 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511867 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511878 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511887 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511895 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.511906 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.511921 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.512228 4945 server.go:940] "Client rotation is on, will bootstrap in background" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.518818 4945 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.518947 4945 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.520891 4945 server.go:997] "Starting client certificate rotation" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.520958 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.521192 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-27 05:41:34.797822083 +0000 UTC Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.521301 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1190h28m23.276525196s for next certificate rotation Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.598323 4945 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.621283 4945 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.683918 4945 log.go:25] "Validated CRI v1 runtime API" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.870133 4945 log.go:25] "Validated CRI v1 image API" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.872667 4945 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.883203 4945 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-08-15-07-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.883262 4945 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.913677 4945 manager.go:217] Machine: {Timestamp:2025-10-08 15:13:11.908675536 +0000 UTC m=+1.262590517 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c21442f4-cc84-4781-8fa8-fec367812c32 BootID:907ff440-3709-4e53-8f2c-af6fb57f351a Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:3e:e9:6f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:3e:e9:6f Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b5:80:51 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:2c:ff:3b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ed:3d:fa Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:fa:19:20 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ce:88:f5:17:f5:7e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:86:6e:a4:34:e7:93 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.914066 4945 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.914312 4945 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.914936 4945 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.915453 4945 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.915519 4945 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.915823 4945 topology_manager.go:138] "Creating topology manager with none policy" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.915841 4945 container_manager_linux.go:303] "Creating device plugin manager" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.918456 4945 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.918507 4945 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.919545 4945 state_mem.go:36] "Initialized new in-memory state store" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.919690 4945 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.925917 4945 kubelet.go:418] "Attempting to sync node with API server" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.925967 4945 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.926071 4945 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.926147 4945 kubelet.go:324] "Adding apiserver pod source" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.926185 4945 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.935014 4945 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.936965 4945 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.938742 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.938844 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:11 crc kubenswrapper[4945]: E1008 15:13:11.938952 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:11 crc kubenswrapper[4945]: E1008 15:13:11.938994 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.944438 4945 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948789 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948846 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948859 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948870 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948892 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948909 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948929 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948959 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.948984 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.949002 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.949022 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.949036 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.949082 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.950219 4945 server.go:1280] "Started kubelet" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.950505 4945 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.950768 4945 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.951754 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.951787 4945 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 08 15:13:11 crc systemd[1]: Started Kubernetes Kubelet. Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.954755 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.954809 4945 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.954953 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 19:49:36.41437226 +0000 UTC Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.955023 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2260h36m24.45935528s for next certificate rotation Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.955089 4945 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 08 15:13:11 crc kubenswrapper[4945]: E1008 15:13:11.955101 4945 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.958725 4945 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.959022 4945 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 08 15:13:11 crc kubenswrapper[4945]: E1008 15:13:11.961319 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="200ms" Oct 08 15:13:11 crc kubenswrapper[4945]: W1008 15:13:11.961577 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:11 crc kubenswrapper[4945]: E1008 15:13:11.961855 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.962800 4945 factory.go:55] Registering systemd factory Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.962944 4945 factory.go:221] Registration of the systemd container factory successfully Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.963508 4945 factory.go:153] Registering CRI-O factory Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.963562 4945 factory.go:221] Registration of the crio container factory successfully Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.963696 4945 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.963731 4945 factory.go:103] Registering Raw factory Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.963763 4945 manager.go:1196] Started watching for new ooms in manager Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.965945 4945 server.go:460] "Adding debug handlers to kubelet server" Oct 08 15:13:11 crc kubenswrapper[4945]: E1008 15:13:11.964305 4945 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c8cd2ed78e150 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 15:13:11.950164304 +0000 UTC m=+1.304079245,LastTimestamp:2025-10-08 15:13:11.950164304 +0000 UTC m=+1.304079245,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.972092 4945 manager.go:319] Starting recovery of all containers Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.976894 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.976951 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.976969 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.976985 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977001 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977015 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977029 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977043 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977059 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977075 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977089 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977131 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977146 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977162 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977177 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977193 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977207 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977221 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977236 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977253 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977268 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977336 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977354 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977371 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977386 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977402 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977424 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977440 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977455 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977472 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977488 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977531 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977548 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977608 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977623 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977640 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977656 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977676 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977693 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977709 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977725 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977740 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977755 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977770 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977785 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977803 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977819 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977834 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977849 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977865 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977879 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977894 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977947 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977972 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.977990 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978007 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978025 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978041 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978058 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978072 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978089 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978104 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978138 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978153 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978168 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978182 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978196 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978211 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978236 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978254 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978270 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978284 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978298 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978314 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978328 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978344 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978359 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978373 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978393 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978411 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978426 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978443 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978460 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978476 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978492 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.978517 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.983915 4945 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984001 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984067 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984088 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984134 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984155 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984178 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984196 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984214 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984231 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984253 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984272 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984290 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984308 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984329 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984359 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984378 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984397 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984419 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984457 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984482 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984508 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984532 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984553 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984571 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984590 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984609 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984627 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984647 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984663 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984680 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984700 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984719 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984735 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984753 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984770 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984786 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984801 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984818 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984838 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984855 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984873 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984893 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984911 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984928 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984946 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984962 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984982 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.984999 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985015 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985033 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985052 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985069 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985085 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985104 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985148 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985165 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985183 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985202 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985222 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985241 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985262 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985280 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985298 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985317 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985336 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985363 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985383 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985417 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985434 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985454 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985474 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985493 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985511 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985528 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985546 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985564 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985582 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985602 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985620 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985642 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985660 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985676 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985693 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985710 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985728 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985746 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985762 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985782 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985800 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985819 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985836 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985853 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985869 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985890 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985909 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985929 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985946 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985963 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.985983 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986000 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986017 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986035 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986053 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986072 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986090 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986132 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986156 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986174 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986192 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986213 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986233 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986263 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986285 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986312 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986333 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986351 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986370 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986390 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986409 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986431 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986451 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986467 4945 reconstruct.go:97] "Volume reconstruction finished" Oct 08 15:13:11 crc kubenswrapper[4945]: I1008 15:13:11.986479 4945 reconciler.go:26] "Reconciler: start to sync state" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.000313 4945 manager.go:324] Recovery completed Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.015210 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.016791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.016835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.016847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.017825 4945 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.017855 4945 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.017881 4945 state_mem.go:36] "Initialized new in-memory state store" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.020754 4945 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.022857 4945 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.022927 4945 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.022960 4945 kubelet.go:2335] "Starting kubelet main sync loop" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.023018 4945 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.023873 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.023960 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.059775 4945 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.123220 4945 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.143427 4945 policy_none.go:49] "None policy: Start" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.144783 4945 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.144815 4945 state_mem.go:35] "Initializing new in-memory state store" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.160788 4945 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.162866 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="400ms" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.213733 4945 manager.go:334] "Starting Device Plugin manager" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.213826 4945 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.213853 4945 server.go:79] "Starting device plugin registration server" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.214611 4945 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.214645 4945 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.214852 4945 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.215308 4945 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.215332 4945 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.226486 4945 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.315138 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.316778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.316840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.316865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.316909 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.317628 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.323815 4945 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.323974 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.325269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.325319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.325336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.325598 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.325896 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.325989 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.326699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.326750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.326772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.326922 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.327067 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.327192 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.327744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.327788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.327807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328288 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328518 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328633 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328693 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.328817 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.329816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.329862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.329883 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.330401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.330449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.330466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.330648 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.330875 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.330931 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.331987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.332037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.332053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.332459 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.332499 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.332687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.332736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.332760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.333752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.333811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.333831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391618 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391654 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391686 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391717 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391866 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391934 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.391973 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.392009 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.392050 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.392081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.392158 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.392209 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.392252 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493634 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493834 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493864 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493892 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493922 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493949 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.493978 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494006 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494035 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494063 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494094 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494439 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494484 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494551 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494587 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494616 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494581 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494596 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494484 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494508 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494688 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494723 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494729 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.494866 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.518343 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.520517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.520579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.520604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.520651 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.521210 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.564339 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="800ms" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.657475 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.665864 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.688102 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.713084 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.721760 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.756044 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.756168 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.815229 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a84ab2c0facd1f99bd4b7477fe5c24490479e1c5cd060f5ac8a03dc9abb2adbb WatchSource:0}: Error finding container a84ab2c0facd1f99bd4b7477fe5c24490479e1c5cd060f5ac8a03dc9abb2adbb: Status 404 returned error can't find the container with id a84ab2c0facd1f99bd4b7477fe5c24490479e1c5cd060f5ac8a03dc9abb2adbb Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.816769 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-cf3e4643347a1b1ac28383a53780abc6848a5fcd763b781a59d415936f0b94c5 WatchSource:0}: Error finding container cf3e4643347a1b1ac28383a53780abc6848a5fcd763b781a59d415936f0b94c5: Status 404 returned error can't find the container with id cf3e4643347a1b1ac28383a53780abc6848a5fcd763b781a59d415936f0b94c5 Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.817968 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-dc25050124535883c969c040af93a9d20ddcc4aff0beeceb590d7f937bbc7503 WatchSource:0}: Error finding container dc25050124535883c969c040af93a9d20ddcc4aff0beeceb590d7f937bbc7503: Status 404 returned error can't find the container with id dc25050124535883c969c040af93a9d20ddcc4aff0beeceb590d7f937bbc7503 Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.818754 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-0e0577c0bd98f332441691bf8d7a27e0a50f6f6c242b9b6d6c764f527da2dbf2 WatchSource:0}: Error finding container 0e0577c0bd98f332441691bf8d7a27e0a50f6f6c242b9b6d6c764f527da2dbf2: Status 404 returned error can't find the container with id 0e0577c0bd98f332441691bf8d7a27e0a50f6f6c242b9b6d6c764f527da2dbf2 Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.819844 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-9b0abd5765732266764309756d67f16e3eb820bcfb9dcc10ccdc15bf9a9cf183 WatchSource:0}: Error finding container 9b0abd5765732266764309756d67f16e3eb820bcfb9dcc10ccdc15bf9a9cf183: Status 404 returned error can't find the container with id 9b0abd5765732266764309756d67f16e3eb820bcfb9dcc10ccdc15bf9a9cf183 Oct 08 15:13:12 crc kubenswrapper[4945]: W1008 15:13:12.905168 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.905256 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.921828 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.923361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.923401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.923411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.923445 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:12 crc kubenswrapper[4945]: E1008 15:13:12.923940 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Oct 08 15:13:12 crc kubenswrapper[4945]: I1008 15:13:12.953551 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.027673 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dc25050124535883c969c040af93a9d20ddcc4aff0beeceb590d7f937bbc7503"} Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.028618 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"cf3e4643347a1b1ac28383a53780abc6848a5fcd763b781a59d415936f0b94c5"} Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.029486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a84ab2c0facd1f99bd4b7477fe5c24490479e1c5cd060f5ac8a03dc9abb2adbb"} Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.030360 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0e0577c0bd98f332441691bf8d7a27e0a50f6f6c242b9b6d6c764f527da2dbf2"} Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.031232 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9b0abd5765732266764309756d67f16e3eb820bcfb9dcc10ccdc15bf9a9cf183"} Oct 08 15:13:13 crc kubenswrapper[4945]: W1008 15:13:13.202383 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:13 crc kubenswrapper[4945]: E1008 15:13:13.202520 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:13 crc kubenswrapper[4945]: E1008 15:13:13.365791 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="1.6s" Oct 08 15:13:13 crc kubenswrapper[4945]: W1008 15:13:13.438419 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:13 crc kubenswrapper[4945]: E1008 15:13:13.438503 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.724719 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.726153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.726257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.726285 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.726336 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:13 crc kubenswrapper[4945]: E1008 15:13:13.727043 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Oct 08 15:13:13 crc kubenswrapper[4945]: I1008 15:13:13.952563 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:14 crc kubenswrapper[4945]: I1008 15:13:14.953319 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:14 crc kubenswrapper[4945]: E1008 15:13:14.966673 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="3.2s" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.037910 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01"} Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.039797 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d" exitCode=0 Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.039961 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.039945 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d"} Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.041629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.041658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.041669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.042316 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc" exitCode=0 Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.042379 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc"} Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.042460 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.043462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.043491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.043502 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.046354 4945 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e7485ffbe62529c9a2f879906870ca3bc5b96d7a2fc704a8c722eb99ddd81bd2" exitCode=0 Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.046442 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.046444 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e7485ffbe62529c9a2f879906870ca3bc5b96d7a2fc704a8c722eb99ddd81bd2"} Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.046594 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.047228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.047270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.047287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.047609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.047642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.047657 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.048965 4945 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d" exitCode=0 Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.049011 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d"} Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.048999 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.051194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.051260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.051287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:15 crc kubenswrapper[4945]: W1008 15:13:15.111084 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:15 crc kubenswrapper[4945]: E1008 15:13:15.111416 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:15 crc kubenswrapper[4945]: W1008 15:13:15.145089 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:15 crc kubenswrapper[4945]: E1008 15:13:15.145242 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:15 crc kubenswrapper[4945]: W1008 15:13:15.256096 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:15 crc kubenswrapper[4945]: E1008 15:13:15.256221 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.327668 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.329213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.329252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.329286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.329323 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:15 crc kubenswrapper[4945]: E1008 15:13:15.329786 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Oct 08 15:13:15 crc kubenswrapper[4945]: W1008 15:13:15.719201 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:15 crc kubenswrapper[4945]: E1008 15:13:15.719317 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Oct 08 15:13:15 crc kubenswrapper[4945]: I1008 15:13:15.952316 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.056184 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.056266 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.056286 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.058729 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8056d279f7555fd7f6ab9350a8550760e8fab4de9d918f9663c2285595d752fd"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.058778 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.060094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.060160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.060176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.061855 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.061900 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.061911 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.061929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.062692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.062732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.062747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.065314 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.065358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.065381 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.065437 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.066842 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.066889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.066904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.067959 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3" exitCode=0 Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.068012 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3"} Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.068149 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.068931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.068973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.068986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:16 crc kubenswrapper[4945]: I1008 15:13:16.952943 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.073172 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd" exitCode=0 Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.073284 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd"} Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.073558 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.075249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.075281 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.075297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.082353 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee"} Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.082406 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.082421 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599"} Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.082405 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.082524 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.082600 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.082621 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.083278 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.083316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.083326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.083865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.083903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.083915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.084040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.084064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.084075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.084153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.084204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.084224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:17 crc kubenswrapper[4945]: I1008 15:13:17.953089 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.088813 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d"} Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.088895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d"} Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.088901 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.088908 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.089072 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.090922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.090997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.091023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.091983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.092036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.092059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:18 crc kubenswrapper[4945]: E1008 15:13:18.168330 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="6.4s" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.530085 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.531807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.531891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.531916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:18 crc kubenswrapper[4945]: I1008 15:13:18.531965 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:18 crc kubenswrapper[4945]: E1008 15:13:18.532922 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.094852 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.097871 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee" exitCode=255 Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.097967 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee"} Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.098187 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.099490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.099534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.099544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.100166 4945 scope.go:117] "RemoveContainer" containerID="8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.106348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528"} Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.106413 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9"} Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.106434 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b"} Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.106818 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.108341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.108412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:19 crc kubenswrapper[4945]: I1008 15:13:19.108439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.112966 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.115905 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf"} Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.116054 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.116101 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.116102 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.117189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.117227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.117241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.117566 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.117644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.117670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.375456 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.978564 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.978852 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.980332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.980376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:20 crc kubenswrapper[4945]: I1008 15:13:20.980389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.047069 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.117884 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.118223 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.118302 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.118350 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.118359 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.119635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.119668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.119682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.119895 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.119952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.120211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.120258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.120281 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.120551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.210155 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:21 crc kubenswrapper[4945]: I1008 15:13:21.674298 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.022015 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.120537 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.120629 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.121478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.121513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.121524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.121788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.121883 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:22 crc kubenswrapper[4945]: I1008 15:13:22.121910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:22 crc kubenswrapper[4945]: E1008 15:13:22.226612 4945 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 15:13:23 crc kubenswrapper[4945]: I1008 15:13:23.122673 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:23 crc kubenswrapper[4945]: I1008 15:13:23.124024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:23 crc kubenswrapper[4945]: I1008 15:13:23.124071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:23 crc kubenswrapper[4945]: I1008 15:13:23.124087 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.118641 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.119187 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.192409 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.192592 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.193845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.193878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.193888 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.201747 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.933386 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.935385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.935562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.935789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:24 crc kubenswrapper[4945]: I1008 15:13:24.935958 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.127504 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.128880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.128937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.128950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.133693 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.551545 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.551776 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.552900 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.552935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:25 crc kubenswrapper[4945]: I1008 15:13:25.552944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:26 crc kubenswrapper[4945]: I1008 15:13:26.130186 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:26 crc kubenswrapper[4945]: I1008 15:13:26.130942 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:26 crc kubenswrapper[4945]: I1008 15:13:26.130974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:26 crc kubenswrapper[4945]: I1008 15:13:26.130987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:28 crc kubenswrapper[4945]: I1008 15:13:28.714827 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 15:13:28 crc kubenswrapper[4945]: I1008 15:13:28.714940 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 15:13:28 crc kubenswrapper[4945]: I1008 15:13:28.721907 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 15:13:28 crc kubenswrapper[4945]: I1008 15:13:28.722032 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.684304 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.685728 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.686357 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.687267 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.687326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.687337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.687306 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 08 15:13:31 crc kubenswrapper[4945]: I1008 15:13:31.693174 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.023260 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.023352 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.148086 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.148660 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.148736 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.149205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.149276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:32 crc kubenswrapper[4945]: I1008 15:13:32.149294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:32 crc kubenswrapper[4945]: E1008 15:13:32.226721 4945 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 15:13:33 crc kubenswrapper[4945]: E1008 15:13:33.699659 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.701575 4945 trace.go:236] Trace[1338216484]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 15:13:20.892) (total time: 12808ms): Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[1338216484]: ---"Objects listed" error: 12808ms (15:13:33.701) Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[1338216484]: [12.808926698s] [12.808926698s] END Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.701637 4945 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.702609 4945 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.704836 4945 trace.go:236] Trace[1506480194]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 15:13:20.300) (total time: 13404ms): Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[1506480194]: ---"Objects listed" error: 13404ms (15:13:33.704) Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[1506480194]: [13.404241325s] [13.404241325s] END Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.705055 4945 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.705091 4945 trace.go:236] Trace[1888356922]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 15:13:19.254) (total time: 14450ms): Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[1888356922]: ---"Objects listed" error: 14450ms (15:13:33.705) Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[1888356922]: [14.450209927s] [14.450209927s] END Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.705332 4945 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.706959 4945 trace.go:236] Trace[149011632]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 15:13:21.446) (total time: 12260ms): Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[149011632]: ---"Objects listed" error: 12260ms (15:13:33.706) Oct 08 15:13:33 crc kubenswrapper[4945]: Trace[149011632]: [12.260836382s] [12.260836382s] END Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.707011 4945 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.805011 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.809228 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.939559 4945 apiserver.go:52] "Watching apiserver" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950015 4945 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950281 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950668 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950770 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950782 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950803 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:33 crc kubenswrapper[4945]: E1008 15:13:33.950918 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.950709 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:33 crc kubenswrapper[4945]: E1008 15:13:33.951127 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:33 crc kubenswrapper[4945]: E1008 15:13:33.951223 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.953142 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.953207 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.957208 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.957287 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.957348 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.957451 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.957568 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.957678 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.958217 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.959613 4945 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.978784 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.987632 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:33 crc kubenswrapper[4945]: I1008 15:13:33.999176 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003154 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003230 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003249 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003271 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003288 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003304 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003322 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003342 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003357 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003375 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003390 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003497 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003531 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003545 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003561 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003577 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003592 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003607 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003624 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003639 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003654 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003670 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003717 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003733 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003748 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003762 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003778 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003794 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003811 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003851 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003885 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003917 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003941 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003957 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003972 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.003987 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004002 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004017 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004032 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004047 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004062 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004077 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004100 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004134 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004152 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004171 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004187 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004202 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004219 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004235 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004252 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004266 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004282 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004298 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004313 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004329 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004345 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004360 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004375 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004402 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004418 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004434 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004451 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004467 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004485 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004499 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004531 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004545 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004559 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004574 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004589 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004606 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004620 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004635 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004651 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004667 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004682 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004697 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004712 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004729 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004744 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004761 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004776 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004792 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004808 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004823 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004843 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004860 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004876 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004897 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004913 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004931 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004951 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004972 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.004997 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005023 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005071 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005095 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005135 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005156 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005198 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005222 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005243 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005259 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005279 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005303 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005326 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005343 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005358 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005407 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005429 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005447 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005468 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005491 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005513 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005534 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005558 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005583 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005607 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005651 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005678 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005701 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005723 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005771 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005793 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005816 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005840 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005861 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005889 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005911 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005935 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005957 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.005981 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006004 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006026 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006048 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006071 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006093 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006134 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006160 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006188 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006238 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006264 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006287 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006310 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006332 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006360 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006410 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006434 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006456 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006478 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006502 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006523 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006547 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006569 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006594 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006621 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006668 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006692 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006715 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006740 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006767 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006791 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006816 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006840 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.006865 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007332 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007361 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007398 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007427 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007436 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007450 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007476 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007502 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007534 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007557 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007581 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007607 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007634 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007658 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007682 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007708 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007733 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007758 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007782 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007834 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007866 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007895 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007922 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007947 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.007997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008023 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008057 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008087 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008130 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008155 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008180 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008202 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.012823 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.016541 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008056 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008301 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008461 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008535 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008635 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.008778 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.009200 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.009652 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.010136 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.010353 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.010651 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.011249 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.011315 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.011499 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.011654 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.011777 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.012323 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.012351 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.012422 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:13:34.512406405 +0000 UTC m=+23.866321306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.018858 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.018896 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.018929 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:34.518911472 +0000 UTC m=+23.872826373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019148 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019263 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019246 4945 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019267 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.012608 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.012724 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.012997 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013099 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013265 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013435 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013400 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013459 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019479 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013511 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013523 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.013754 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.015086 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.015338 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.015395 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.015555 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.015760 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.015861 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.016035 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.016137 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.016259 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.016274 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.016859 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.017019 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.020353 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:34.520328678 +0000 UTC m=+23.874243759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.020974 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.020999 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.020033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019420 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019685 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.012573 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019726 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019746 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019928 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019951 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.019995 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.020024 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.020041 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.020088 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021284 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021323 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021481 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021673 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021808 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.021873 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.022052 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.022147 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.022803 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.022921 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.023098 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.023095 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.023417 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.023730 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.023941 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.023957 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.024263 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.024721 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.024796 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.024973 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.025027 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.025219 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.025483 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.025522 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.025710 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.025905 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.025919 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.026061 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.026248 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.026316 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.026531 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.026670 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.026982 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.027027 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.027049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.027462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.027612 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.027661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.027738 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.027949 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.028004 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.028000 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.028202 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.028343 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.028358 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.028600 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.028732 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.029003 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.029425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.029519 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.029847 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.029932 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030046 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030122 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030377 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030395 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030434 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030520 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030505 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030615 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030629 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030622 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030714 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030766 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.030824 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032150 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.031382 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.031533 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.031588 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.031547 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.031820 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.031253 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032321 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032587 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032606 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032696 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032904 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032937 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.032946 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033186 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033207 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033349 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033416 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033448 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033454 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033582 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.034096 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.033649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.034398 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.034628 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035053 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035097 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035383 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035618 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035773 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035779 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035796 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035834 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035911 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.035925 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.036012 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.036414 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.036620 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.036838 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.036899 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.037144 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.037500 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.037588 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.037606 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.037709 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.037839 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.037666 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:34.537646323 +0000 UTC m=+23.891561454 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038082 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038228 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038253 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038668 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038690 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038712 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038788 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038791 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038850 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038884 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038952 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038963 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.038994 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.039320 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.039333 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.039669 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.039686 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.040037 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.040787 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.041285 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.041314 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.041312 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.041329 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.041387 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:34.541367939 +0000 UTC m=+23.895283050 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.045399 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.045518 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.045544 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.045702 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.046167 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.046204 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.055730 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.056375 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.057575 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.059752 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.061536 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.063828 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.064876 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.067005 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.067056 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.070050 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.070159 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.071877 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.073185 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.074646 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.075433 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.076411 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.076729 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.077701 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.078935 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.079815 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.080492 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.080510 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.081479 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.082204 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.083163 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.083884 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.084616 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.085692 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.086419 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.087500 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.088361 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.088937 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.090249 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.090497 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.091218 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.092345 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.093164 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.094201 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.094822 4945 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.094954 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.097694 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.098192 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.098587 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.099838 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.100222 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.101206 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.101759 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.102803 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.103459 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.104287 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.104846 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.105855 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.107263 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.107936 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109269 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109430 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109496 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109592 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109704 4945 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109729 4945 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109742 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109754 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109766 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109777 4945 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109788 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109800 4945 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.109883 4945 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110057 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110422 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110445 4945 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110491 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110512 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110529 4945 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110545 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110562 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110579 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110596 4945 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110613 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110630 4945 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110645 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110662 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110679 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110695 4945 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110710 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110726 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110741 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110714 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110756 4945 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110778 4945 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110786 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110795 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110803 4945 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110812 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110821 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110829 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110838 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110952 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110975 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.110987 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111000 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111012 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111024 4945 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111042 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111054 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111066 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111079 4945 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111090 4945 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111102 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111148 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111161 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111173 4945 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111184 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111196 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111207 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111218 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111230 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111242 4945 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111255 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111266 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111277 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111288 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111299 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111310 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111325 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111337 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111350 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111361 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111372 4945 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111385 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111398 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111410 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111422 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111432 4945 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111444 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111456 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111468 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111480 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111491 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111502 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111515 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111526 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111537 4945 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111547 4945 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111559 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111570 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111581 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111592 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111604 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111615 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111627 4945 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111639 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111651 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111663 4945 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111676 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111688 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111691 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111698 4945 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111822 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111837 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111850 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111862 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111875 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111887 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111899 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111910 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111922 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111934 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111944 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111955 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111970 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.111986 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112003 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112019 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112031 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112042 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112052 4945 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112086 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112099 4945 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112135 4945 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112152 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112165 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112176 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112187 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112201 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112217 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112233 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112246 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112257 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112268 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112280 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112291 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112302 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112314 4945 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112326 4945 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112337 4945 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112348 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112359 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112371 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112383 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112388 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112394 4945 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112519 4945 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112531 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112543 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112554 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112565 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112577 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112589 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112600 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112612 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112624 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112636 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112647 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112659 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112670 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112681 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112694 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112706 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112717 4945 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112729 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112740 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112752 4945 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112764 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112776 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112788 4945 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112799 4945 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112811 4945 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112822 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112834 4945 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112846 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112886 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112899 4945 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112911 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112923 4945 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112935 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112952 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112964 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112975 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112987 4945 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.112999 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113011 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113023 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113035 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113046 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113058 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113070 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113082 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113143 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113045 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113179 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113223 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113237 4945 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113250 4945 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113261 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113273 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113285 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113297 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.113308 4945 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.114372 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.115096 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.116540 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.117021 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.158692 4945 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.199432 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58064->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.199491 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58064->192.168.126.11:17697: read: connection reset by peer" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.265233 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.282895 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.296993 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.519465 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.519896 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.520028 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.520085 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:35.520069001 +0000 UTC m=+24.873983902 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.520574 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:13:35.520563525 +0000 UTC m=+24.874478416 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.621320 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.621439 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:34 crc kubenswrapper[4945]: I1008 15:13:34.621472 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621583 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621618 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621633 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621658 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621685 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621660 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621698 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621704 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:35.621683991 +0000 UTC m=+24.975598892 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621907 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:35.621886796 +0000 UTC m=+24.975801687 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:34 crc kubenswrapper[4945]: E1008 15:13:34.621933 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:35.621925757 +0000 UTC m=+24.975840658 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.158539 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.159168 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.160635 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf" exitCode=255 Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.160716 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf"} Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.160780 4945 scope.go:117] "RemoveContainer" containerID="8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.162477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4"} Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.162524 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ca42a15fdc4282dffea2dec2af159a00925d6d080904c5d2a8ddff7b39dc37a6"} Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.163910 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade"} Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.163946 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98"} Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.163958 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6073588ae8efc7ca58e374c44d2fa320632d3db4acc2d5b28b20c142389d36dc"} Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.165195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"92abc1d1abb6d978fa243c74dae2029c6120f18bdadc6e91e42fafe959352c96"} Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.173543 4945 scope.go:117] "RemoveContainer" containerID="9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.173688 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.173826 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.173798 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.187549 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.197846 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.208972 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.219877 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.230714 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.243489 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.251573 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.260437 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.278446 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.289723 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.302126 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.314621 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.330549 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:18Z\\\",\\\"message\\\":\\\"W1008 15:13:17.123992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 15:13:17.125032 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759936397 cert, and key in /tmp/serving-cert-3301865878/serving-signer.crt, /tmp/serving-cert-3301865878/serving-signer.key\\\\nI1008 15:13:17.406749 1 observer_polling.go:159] Starting file observer\\\\nW1008 15:13:17.413984 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 15:13:17.414268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:17.422143 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3301865878/tls.crt::/tmp/serving-cert-3301865878/tls.key\\\\\\\"\\\\nF1008 15:13:17.991500 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.342808 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.529684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.529769 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.529851 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.529871 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:13:37.529832724 +0000 UTC m=+26.883747645 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.529948 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:37.529900016 +0000 UTC m=+26.883815147 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.580577 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.593315 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.595507 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.601532 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.609083 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.616718 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.625511 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:18Z\\\",\\\"message\\\":\\\"W1008 15:13:17.123992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 15:13:17.125032 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759936397 cert, and key in /tmp/serving-cert-3301865878/serving-signer.crt, /tmp/serving-cert-3301865878/serving-signer.key\\\\nI1008 15:13:17.406749 1 observer_polling.go:159] Starting file observer\\\\nW1008 15:13:17.413984 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 15:13:17.414268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:17.422143 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3301865878/tls.crt::/tmp/serving-cert-3301865878/tls.key\\\\\\\"\\\\nF1008 15:13:17.991500 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.627968 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.630930 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.630969 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.630990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631122 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631131 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631142 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631160 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631139 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631192 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:37.631176086 +0000 UTC m=+26.985091157 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631207 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:37.631201207 +0000 UTC m=+26.985116108 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631216 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631232 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:35 crc kubenswrapper[4945]: E1008 15:13:35.631275 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:37.631261058 +0000 UTC m=+26.985176139 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.635159 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.642665 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.650803 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.658554 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.669953 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.682740 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.711745 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.724026 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8abc91500571edf9e6423b7a511a2bac2998d0548e705865e5de57bb96b605ee\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:18Z\\\",\\\"message\\\":\\\"W1008 15:13:17.123992 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 15:13:17.125032 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759936397 cert, and key in /tmp/serving-cert-3301865878/serving-signer.crt, /tmp/serving-cert-3301865878/serving-signer.key\\\\nI1008 15:13:17.406749 1 observer_polling.go:159] Starting file observer\\\\nW1008 15:13:17.413984 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 15:13:17.414268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:17.422143 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3301865878/tls.crt::/tmp/serving-cert-3301865878/tls.key\\\\\\\"\\\\nF1008 15:13:17.991500 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.736592 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.747525 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.760481 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:35 crc kubenswrapper[4945]: I1008 15:13:35.771688 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:35Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.023763 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.023866 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:36 crc kubenswrapper[4945]: E1008 15:13:36.023950 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.023879 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:36 crc kubenswrapper[4945]: E1008 15:13:36.024076 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:36 crc kubenswrapper[4945]: E1008 15:13:36.024279 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.168379 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.170773 4945 scope.go:117] "RemoveContainer" containerID="9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf" Oct 08 15:13:36 crc kubenswrapper[4945]: E1008 15:13:36.170903 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 08 15:13:36 crc kubenswrapper[4945]: E1008 15:13:36.187756 4945 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.189756 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.203354 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.217433 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.228590 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.245953 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.264703 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.276466 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.288984 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:36 crc kubenswrapper[4945]: I1008 15:13:36.301428 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:36Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.173763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708"} Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.189958 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.205232 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.222022 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.234865 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.244883 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.254681 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.272191 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.284855 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.295560 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:37Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.552380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.552574 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:13:41.552543329 +0000 UTC m=+30.906458270 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.552693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.552789 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.552843 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:41.552833316 +0000 UTC m=+30.906748427 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.653419 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.653473 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:37 crc kubenswrapper[4945]: I1008 15:13:37.653504 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653649 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653670 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653684 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653700 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653744 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653749 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653766 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653747 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:41.653728288 +0000 UTC m=+31.007643209 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653925 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:41.653899792 +0000 UTC m=+31.007814723 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:37 crc kubenswrapper[4945]: E1008 15:13:37.653951 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:41.653936023 +0000 UTC m=+31.007850964 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:38 crc kubenswrapper[4945]: I1008 15:13:38.023812 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:38 crc kubenswrapper[4945]: I1008 15:13:38.023890 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:38 crc kubenswrapper[4945]: I1008 15:13:38.023888 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:38 crc kubenswrapper[4945]: E1008 15:13:38.024008 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:38 crc kubenswrapper[4945]: E1008 15:13:38.024135 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:38 crc kubenswrapper[4945]: E1008 15:13:38.024213 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.023615 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.023649 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.023615 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.023757 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.023880 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.023944 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.389727 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-z7jwc"] Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.390063 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.392188 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.392376 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.392399 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.401600 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.415542 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.428901 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.442016 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.455542 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.477136 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7a1789a-2045-4ec7-884e-eb70c9031197-hosts-file\") pod \"node-resolver-z7jwc\" (UID: \"d7a1789a-2045-4ec7-884e-eb70c9031197\") " pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.477184 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xq6k\" (UniqueName: \"kubernetes.io/projected/d7a1789a-2045-4ec7-884e-eb70c9031197-kube-api-access-5xq6k\") pod \"node-resolver-z7jwc\" (UID: \"d7a1789a-2045-4ec7-884e-eb70c9031197\") " pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.477671 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.491968 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.504936 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.517003 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.531429 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.577927 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7a1789a-2045-4ec7-884e-eb70c9031197-hosts-file\") pod \"node-resolver-z7jwc\" (UID: \"d7a1789a-2045-4ec7-884e-eb70c9031197\") " pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.577979 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xq6k\" (UniqueName: \"kubernetes.io/projected/d7a1789a-2045-4ec7-884e-eb70c9031197-kube-api-access-5xq6k\") pod \"node-resolver-z7jwc\" (UID: \"d7a1789a-2045-4ec7-884e-eb70c9031197\") " pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.578158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7a1789a-2045-4ec7-884e-eb70c9031197-hosts-file\") pod \"node-resolver-z7jwc\" (UID: \"d7a1789a-2045-4ec7-884e-eb70c9031197\") " pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.601392 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xq6k\" (UniqueName: \"kubernetes.io/projected/d7a1789a-2045-4ec7-884e-eb70c9031197-kube-api-access-5xq6k\") pod \"node-resolver-z7jwc\" (UID: \"d7a1789a-2045-4ec7-884e-eb70c9031197\") " pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.700347 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.701726 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z7jwc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.702061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.702096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.702125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.702189 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 15:13:40 crc kubenswrapper[4945]: W1008 15:13:40.721420 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7a1789a_2045_4ec7_884e_eb70c9031197.slice/crio-56439b89f7cef111b500274e7f53ea9f2b8d4cfe0d744c27d99eafeade755ba1 WatchSource:0}: Error finding container 56439b89f7cef111b500274e7f53ea9f2b8d4cfe0d744c27d99eafeade755ba1: Status 404 returned error can't find the container with id 56439b89f7cef111b500274e7f53ea9f2b8d4cfe0d744c27d99eafeade755ba1 Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.724127 4945 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.724403 4945 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.725257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.725286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.725298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.725316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.725328 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:40Z","lastTransitionTime":"2025-10-08T15:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.751392 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-jp64g"] Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.751789 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.753574 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.753816 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.754193 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.754422 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.754911 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.755359 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.760756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.760796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.760805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.760822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.760831 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:40Z","lastTransitionTime":"2025-10-08T15:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.804136 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.836223 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.847147 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.847572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.847602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.847613 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.847629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.847643 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:40Z","lastTransitionTime":"2025-10-08T15:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.880680 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njl6g\" (UniqueName: \"kubernetes.io/projected/e39fc950-7f86-4d6b-b4b0-25daafb66250-kube-api-access-njl6g\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.880745 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e39fc950-7f86-4d6b-b4b0-25daafb66250-rootfs\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.880773 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e39fc950-7f86-4d6b-b4b0-25daafb66250-proxy-tls\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.880811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e39fc950-7f86-4d6b-b4b0-25daafb66250-mcd-auth-proxy-config\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.882328 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.888489 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.892424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.892463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.892474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.892493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.892506 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:40Z","lastTransitionTime":"2025-10-08T15:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.918547 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.924590 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.928850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.928904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.928917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.928932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.928941 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:40Z","lastTransitionTime":"2025-10-08T15:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.935889 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.939778 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: E1008 15:13:40.939921 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.941558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.941605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.941616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.941635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.941647 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:40Z","lastTransitionTime":"2025-10-08T15:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.948385 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.957800 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.972274 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.982090 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e39fc950-7f86-4d6b-b4b0-25daafb66250-rootfs\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.982152 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e39fc950-7f86-4d6b-b4b0-25daafb66250-proxy-tls\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.982180 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e39fc950-7f86-4d6b-b4b0-25daafb66250-mcd-auth-proxy-config\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.982182 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e39fc950-7f86-4d6b-b4b0-25daafb66250-rootfs\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.982302 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njl6g\" (UniqueName: \"kubernetes.io/projected/e39fc950-7f86-4d6b-b4b0-25daafb66250-kube-api-access-njl6g\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.982974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e39fc950-7f86-4d6b-b4b0-25daafb66250-mcd-auth-proxy-config\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.988807 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e39fc950-7f86-4d6b-b4b0-25daafb66250-proxy-tls\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:40 crc kubenswrapper[4945]: I1008 15:13:40.991896 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:40Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.008151 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.009548 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njl6g\" (UniqueName: \"kubernetes.io/projected/e39fc950-7f86-4d6b-b4b0-25daafb66250-kube-api-access-njl6g\") pod \"machine-config-daemon-jp64g\" (UID: \"e39fc950-7f86-4d6b-b4b0-25daafb66250\") " pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.025780 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.044100 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.044152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.044162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.044180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.044191 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.065527 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:13:41 crc kubenswrapper[4945]: W1008 15:13:41.077231 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode39fc950_7f86_4d6b_b4b0_25daafb66250.slice/crio-ae1a01cb78e6c689c0478c859020b012e0d681ffacac25ce6f9bac78c0a24c59 WatchSource:0}: Error finding container ae1a01cb78e6c689c0478c859020b012e0d681ffacac25ce6f9bac78c0a24c59: Status 404 returned error can't find the container with id ae1a01cb78e6c689c0478c859020b012e0d681ffacac25ce6f9bac78c0a24c59 Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.146451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.146487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.146497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.146516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.146529 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.156833 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-j8kc9"] Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.157291 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2gdgm"] Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.157524 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.157981 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.159658 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.160321 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.160624 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.160850 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.160882 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.160988 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.161160 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.178366 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.186455 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"ae1a01cb78e6c689c0478c859020b012e0d681ffacac25ce6f9bac78c0a24c59"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.187790 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z7jwc" event={"ID":"d7a1789a-2045-4ec7-884e-eb70c9031197","Type":"ContainerStarted","Data":"fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.187832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z7jwc" event={"ID":"d7a1789a-2045-4ec7-884e-eb70c9031197","Type":"ContainerStarted","Data":"56439b89f7cef111b500274e7f53ea9f2b8d4cfe0d744c27d99eafeade755ba1"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.189647 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.203452 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.221718 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.236655 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.248666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.248728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.248738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.248751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.248767 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.249614 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.263574 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.277647 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.284639 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.284731 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-netns\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.284812 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-os-release\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.284946 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-socket-dir-parent\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.285011 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-multus-certs\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.285252 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-system-cni-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.285433 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-os-release\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.285563 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-cni-binary-copy\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.285670 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-conf-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.285836 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-etc-kubernetes\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.285979 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-cni-multus\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.286184 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-kubelet\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.286365 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mzlm\" (UniqueName: \"kubernetes.io/projected/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-kube-api-access-4mzlm\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.286529 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-cni-bin\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.286673 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-hostroot\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.286829 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-cni-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.286965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-k8s-cni-cncf-io\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.287171 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-daemon-config\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.287297 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.287353 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rksk\" (UniqueName: \"kubernetes.io/projected/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-kube-api-access-2rksk\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.287465 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cnibin\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.287537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-system-cni-dir\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.287585 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cni-binary-copy\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.287724 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-cnibin\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.288040 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.299370 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.313097 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.333979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.350654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.350698 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.350714 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.350736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.350751 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.352307 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.367884 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.381274 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388502 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-cnibin\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388549 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388577 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-netns\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388609 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-os-release\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388639 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-socket-dir-parent\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388662 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-multus-certs\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388662 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-cnibin\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-system-cni-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388718 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-os-release\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388742 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-cni-binary-copy\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388764 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-conf-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-etc-kubernetes\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388797 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-cni-multus\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-netns\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-kubelet\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388814 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-kubelet\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388822 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-multus-certs\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388925 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mzlm\" (UniqueName: \"kubernetes.io/projected/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-kube-api-access-4mzlm\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388968 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-cni-multus\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388981 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-system-cni-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388982 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-socket-dir-parent\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-cni-bin\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389042 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-hostroot\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389052 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-os-release\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389141 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-var-lib-cni-bin\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-cni-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389074 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-os-release\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389196 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-conf-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389200 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-hostroot\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389261 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-k8s-cni-cncf-io\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389283 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-daemon-config\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389300 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389317 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-cni-dir\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389329 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rksk\" (UniqueName: \"kubernetes.io/projected/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-kube-api-access-2rksk\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389346 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-host-run-k8s-cni-cncf-io\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389394 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cnibin\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389360 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cnibin\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389463 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-system-cni-dir\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389504 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cni-binary-copy\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389616 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-cni-binary-copy\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.389609 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-system-cni-dir\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.388998 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-etc-kubernetes\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.390024 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-multus-daemon-config\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.390151 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.390477 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-cni-binary-copy\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.392515 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.393369 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.405743 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mzlm\" (UniqueName: \"kubernetes.io/projected/e93c84e1-87ae-4485-8ea1-8cd4ac1554a4-kube-api-access-4mzlm\") pod \"multus-additional-cni-plugins-2gdgm\" (UID: \"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\") " pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.406831 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rksk\" (UniqueName: \"kubernetes.io/projected/22e38fb8-8588-494f-a6cb-8b3d2c3a3142-kube-api-access-2rksk\") pod \"multus-j8kc9\" (UID: \"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\") " pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.406909 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.421415 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.435605 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.447662 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.452940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.452968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.452976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.452990 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.453000 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.459213 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.471905 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.478146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-j8kc9" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.481408 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.485430 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: W1008 15:13:41.488559 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22e38fb8_8588_494f_a6cb_8b3d2c3a3142.slice/crio-e07cd400862785221c27c0d7b6ff1f468b325a878b886b7593e28686df87eb2c WatchSource:0}: Error finding container e07cd400862785221c27c0d7b6ff1f468b325a878b886b7593e28686df87eb2c: Status 404 returned error can't find the container with id e07cd400862785221c27c0d7b6ff1f468b325a878b886b7593e28686df87eb2c Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.507925 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.520092 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.537441 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9g2vg"] Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.538557 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.540824 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.540996 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.541231 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.541461 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.541591 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.542681 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.542967 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.553456 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.555661 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.555708 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.555720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.555735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.555749 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.572006 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.589387 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.590553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.590697 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.590732 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:13:49.590709254 +0000 UTC m=+38.944624345 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.590789 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.590857 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:49.590837537 +0000 UTC m=+38.944752638 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.602099 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.617324 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.635839 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.655920 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.658422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.658469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.658481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.658499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.658512 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.670769 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.682432 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692256 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-netns\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692295 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-ovn-kubernetes\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692316 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-kubelet\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692336 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-config\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-ovn\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692458 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-node-log\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692481 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-netd\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692532 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-script-lib\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692595 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692632 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692656 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-etc-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.692662 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692680 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.692704 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:49.692690563 +0000 UTC m=+39.046605464 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fzlp\" (UniqueName: \"kubernetes.io/projected/28239584-598a-49d2-a9b0-189e4f4ad733-kube-api-access-8fzlp\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-systemd\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692790 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-env-overrides\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.692791 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.692812 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.692825 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692825 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-slash\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692850 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-log-socket\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692871 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-var-lib-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.692893 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:49.692868848 +0000 UTC m=+39.046783749 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692926 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692955 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-systemd-units\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.692978 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-bin\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.693000 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28239584-598a-49d2-a9b0-189e4f4ad733-ovn-node-metrics-cert\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.693028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.693155 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.693174 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.693184 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:41 crc kubenswrapper[4945]: E1008 15:13:41.693217 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:49.693206867 +0000 UTC m=+39.047121968 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.694452 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.705227 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.715821 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.726550 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.737145 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:41Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.761503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.761538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.761547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.761561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.761572 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794158 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-systemd-units\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794208 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-var-lib-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794251 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-bin\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794271 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794291 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28239584-598a-49d2-a9b0-189e4f4ad733-ovn-node-metrics-cert\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794316 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-netns\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-ovn-kubernetes\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794366 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-kubelet\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794387 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-node-log\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-systemd-units\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-var-lib-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794497 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-netns\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794521 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-netd\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794411 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-bin\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794524 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-kubelet\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794497 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-ovn-kubernetes\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794548 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-node-log\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-netd\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794586 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794639 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-config\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794692 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-ovn\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794731 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-script-lib\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794784 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-ovn\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794809 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-etc-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794843 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794876 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fzlp\" (UniqueName: \"kubernetes.io/projected/28239584-598a-49d2-a9b0-189e4f4ad733-kube-api-access-8fzlp\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794907 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-systemd\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.794972 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-env-overrides\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-slash\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795046 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-log-socket\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795157 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-log-socket\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795228 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-etc-openvswitch\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795252 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-systemd\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795266 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795333 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-config\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795335 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-slash\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.795485 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-script-lib\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.796006 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-env-overrides\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.801544 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28239584-598a-49d2-a9b0-189e4f4ad733-ovn-node-metrics-cert\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.821000 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fzlp\" (UniqueName: \"kubernetes.io/projected/28239584-598a-49d2-a9b0-189e4f4ad733-kube-api-access-8fzlp\") pod \"ovnkube-node-9g2vg\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.849399 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.864419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.864482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.864504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.864534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.864556 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:41 crc kubenswrapper[4945]: W1008 15:13:41.964408 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28239584_598a_49d2_a9b0_189e4f4ad733.slice/crio-6ec74449b25519e4606575db149042f9c896c0632effe38a06596c77fc39fa9b WatchSource:0}: Error finding container 6ec74449b25519e4606575db149042f9c896c0632effe38a06596c77fc39fa9b: Status 404 returned error can't find the container with id 6ec74449b25519e4606575db149042f9c896c0632effe38a06596c77fc39fa9b Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.966490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.966534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.966545 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.966562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:41 crc kubenswrapper[4945]: I1008 15:13:41.966573 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:41Z","lastTransitionTime":"2025-10-08T15:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.024022 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.024077 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.024178 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:42 crc kubenswrapper[4945]: E1008 15:13:42.024187 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:42 crc kubenswrapper[4945]: E1008 15:13:42.024324 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:42 crc kubenswrapper[4945]: E1008 15:13:42.024482 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.043478 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.054782 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.065180 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.069101 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.069154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.069163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.069179 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.069188 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.076611 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.087917 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.099996 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.114027 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.126136 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.142043 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.154581 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.167175 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.171628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.171691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.171704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.171724 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.171744 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.179889 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.191443 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"6ec74449b25519e4606575db149042f9c896c0632effe38a06596c77fc39fa9b"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.192809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerStarted","Data":"88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.193003 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerStarted","Data":"ce257190187154d66d679f7f24d4b403d4ecf9fb1ba3d8fcf387930a2f4e193f"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.194043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerStarted","Data":"e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.194067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerStarted","Data":"e07cd400862785221c27c0d7b6ff1f468b325a878b886b7593e28686df87eb2c"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.194508 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.196020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.196064 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.214457 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.231565 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.249900 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.268052 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.274203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.274243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.274252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.274268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.274279 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.281796 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.292342 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.305687 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.322548 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.341176 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.355330 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.376092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.376178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.376190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.376207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.376219 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.384759 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.409715 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.424568 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.463886 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.477942 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.477980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.477989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.478007 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.478021 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.499406 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.580946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.581014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.581031 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.581055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.581071 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.683674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.683707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.683716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.683730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.683739 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.786350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.786400 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.786412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.786429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.786443 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.889145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.889210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.889228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.889251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.889268 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.992704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.992773 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.992790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.993011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:42 crc kubenswrapper[4945]: I1008 15:13:42.993400 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:42Z","lastTransitionTime":"2025-10-08T15:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.096371 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.096413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.096423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.096437 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.096446 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.198536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.198575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.198587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.198607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.198620 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.200494 4945 generic.go:334] "Generic (PLEG): container finished" podID="e93c84e1-87ae-4485-8ea1-8cd4ac1554a4" containerID="88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259" exitCode=0 Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.200540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerDied","Data":"88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.202010 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d" exitCode=0 Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.202081 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.217757 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.241153 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.252963 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.264664 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.281345 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.298223 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.302690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.302723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.302734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.302749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.302760 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.316976 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.336386 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.349695 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.363718 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.374734 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.389222 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.400350 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.404494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.404524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.404535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.404549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.404558 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.413036 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.424150 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.436824 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.449447 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.471755 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.482495 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.492828 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.505236 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.506754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.506789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.506803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.506819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.506832 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.519248 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.537231 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.550570 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.562525 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.574491 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.586644 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.614180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.614232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.614243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.614262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.614273 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.621683 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.716284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.716322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.716331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.716346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.716357 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.725644 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-qb4pj"] Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.726168 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.728449 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.728515 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.728683 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.729624 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.740039 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.779473 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.814059 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4138a218-2d22-4042-ae74-5d4712982025-serviceca\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.814136 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4138a218-2d22-4042-ae74-5d4712982025-host\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.814355 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkfnq\" (UniqueName: \"kubernetes.io/projected/4138a218-2d22-4042-ae74-5d4712982025-kube-api-access-gkfnq\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.818801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.818845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.818857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.818877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.818889 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.829919 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.859682 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.872267 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.873043 4945 scope.go:117] "RemoveContainer" containerID="9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf" Oct 08 15:13:43 crc kubenswrapper[4945]: E1008 15:13:43.873213 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.900618 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.914997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4138a218-2d22-4042-ae74-5d4712982025-serviceca\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.915179 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4138a218-2d22-4042-ae74-5d4712982025-host\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.915260 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4138a218-2d22-4042-ae74-5d4712982025-host\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.915357 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkfnq\" (UniqueName: \"kubernetes.io/projected/4138a218-2d22-4042-ae74-5d4712982025-kube-api-access-gkfnq\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.920985 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4138a218-2d22-4042-ae74-5d4712982025-serviceca\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.921824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.921846 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.921855 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.921869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.921880 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:43Z","lastTransitionTime":"2025-10-08T15:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.938631 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:43Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:43 crc kubenswrapper[4945]: I1008 15:13:43.970535 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkfnq\" (UniqueName: \"kubernetes.io/projected/4138a218-2d22-4042-ae74-5d4712982025-kube-api-access-gkfnq\") pod \"node-ca-qb4pj\" (UID: \"4138a218-2d22-4042-ae74-5d4712982025\") " pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.005479 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.023186 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.023276 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.023292 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:44 crc kubenswrapper[4945]: E1008 15:13:44.023350 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:44 crc kubenswrapper[4945]: E1008 15:13:44.023445 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:44 crc kubenswrapper[4945]: E1008 15:13:44.023582 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.024393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.024457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.024480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.024507 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.024530 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.041052 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qb4pj" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.057876 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: W1008 15:13:44.062564 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4138a218_2d22_4042_ae74_5d4712982025.slice/crio-0ed64ff3d502e7fc7526a0bdd683bc142d69c11b741f147d887ae9a6df1fe281 WatchSource:0}: Error finding container 0ed64ff3d502e7fc7526a0bdd683bc142d69c11b741f147d887ae9a6df1fe281: Status 404 returned error can't find the container with id 0ed64ff3d502e7fc7526a0bdd683bc142d69c11b741f147d887ae9a6df1fe281 Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.099869 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.135844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.136262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.136287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.136311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.136330 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.135878 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.162941 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.201834 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.206763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qb4pj" event={"ID":"4138a218-2d22-4042-ae74-5d4712982025","Type":"ContainerStarted","Data":"0ed64ff3d502e7fc7526a0bdd683bc142d69c11b741f147d887ae9a6df1fe281"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.211973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.212023 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.212036 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.216069 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerStarted","Data":"c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.238238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.238280 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.238289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.239191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.239259 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.245506 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.281675 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.322496 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.341404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.341474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.341499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.341529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.341551 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.361069 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.400987 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.444230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.444270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.444290 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.444309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.444321 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.447731 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.498402 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.524832 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.546368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.546405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.546418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.546436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.546447 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.559274 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.604638 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.644730 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.648422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.648463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.648476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.648494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.648506 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.683719 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.720632 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.751056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.751174 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.751191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.751215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.751227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.762575 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.803640 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.844952 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.853084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.853140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.853158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.853182 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.853198 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.884771 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.920632 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:44Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.955878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.955938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.955955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.955985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:44 crc kubenswrapper[4945]: I1008 15:13:44.956002 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:44Z","lastTransitionTime":"2025-10-08T15:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.059415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.059452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.059462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.059478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.059491 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.162325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.162758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.162776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.162803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.162822 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.221592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qb4pj" event={"ID":"4138a218-2d22-4042-ae74-5d4712982025","Type":"ContainerStarted","Data":"a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.226292 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.226382 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.226408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.229480 4945 generic.go:334] "Generic (PLEG): container finished" podID="e93c84e1-87ae-4485-8ea1-8cd4ac1554a4" containerID="c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d" exitCode=0 Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.229551 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerDied","Data":"c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.242749 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.260286 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.267970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.268033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.268044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.268063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.268079 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.280019 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.305777 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.327907 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.340759 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.352531 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.367273 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.371325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.371357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.371366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.371381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.371391 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.380437 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.392883 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.402866 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.412711 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.441942 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.474413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.474459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.474470 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.474489 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.474503 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.483974 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.520985 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.560025 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.577424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.577465 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.577478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.577495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.577509 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.603140 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.646591 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.680159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.680203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.680216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.680236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.680253 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.686434 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.725654 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.761614 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.783072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.783192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.783213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.783242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.783264 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.808458 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.842964 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.886277 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.886337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.886356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.886385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.886407 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.888702 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.920928 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.966469 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.988985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.989022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.989033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.989048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:45 crc kubenswrapper[4945]: I1008 15:13:45.989060 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:45Z","lastTransitionTime":"2025-10-08T15:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.002982 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.024156 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.024434 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:46 crc kubenswrapper[4945]: E1008 15:13:46.024517 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:46 crc kubenswrapper[4945]: E1008 15:13:46.024579 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.024434 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:46 crc kubenswrapper[4945]: E1008 15:13:46.024941 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.046941 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.089092 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.091528 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.091594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.091613 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.091634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.091653 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.123764 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.194693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.195282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.195294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.195314 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.195329 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.235690 4945 generic.go:334] "Generic (PLEG): container finished" podID="e93c84e1-87ae-4485-8ea1-8cd4ac1554a4" containerID="6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2" exitCode=0 Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.235790 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerDied","Data":"6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.260924 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.277571 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.292563 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.298248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.298302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.298315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.298341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.298355 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.308185 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.323517 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.366713 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.402012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.402078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.402092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.402137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.402156 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.403219 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.438179 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.480730 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.505259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.505321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.505336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.505356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.505425 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.523588 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.566258 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.602562 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.608380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.608445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.608463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.608490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.608507 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.639927 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.680249 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.711089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.711158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.711172 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.711188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.711200 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.724557 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.813982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.814025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.814038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.814056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.814093 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.917262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.917297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.917306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.917319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:46 crc kubenswrapper[4945]: I1008 15:13:46.917331 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:46Z","lastTransitionTime":"2025-10-08T15:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.019857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.019907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.019925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.019948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.019965 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.122336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.122390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.122408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.122432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.122451 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.225843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.225919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.225951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.225985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.226014 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.246604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.250768 4945 generic.go:334] "Generic (PLEG): container finished" podID="e93c84e1-87ae-4485-8ea1-8cd4ac1554a4" containerID="d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc" exitCode=0 Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.250820 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerDied","Data":"d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.277795 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.298867 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.314917 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.328773 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.328819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.328831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.328848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.328861 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.331509 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.352647 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.377526 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.394536 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.411541 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.424461 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.431480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.431524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.431541 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.431564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.431580 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.442778 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.457536 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.470713 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.485270 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.494460 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.506007 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:47Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.533832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.533864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.533874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.533887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.533898 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.636754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.636802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.636815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.636832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.636847 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.739062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.739184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.739210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.739239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.739260 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.842174 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.842289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.842318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.842344 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.842362 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.947643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.947717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.947738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.947764 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:47 crc kubenswrapper[4945]: I1008 15:13:47.947784 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:47Z","lastTransitionTime":"2025-10-08T15:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.023347 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.023495 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:48 crc kubenswrapper[4945]: E1008 15:13:48.023575 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.023649 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:48 crc kubenswrapper[4945]: E1008 15:13:48.023757 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:48 crc kubenswrapper[4945]: E1008 15:13:48.023907 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.050850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.050911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.050937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.050959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.050972 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.155078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.155212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.155233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.155309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.155328 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.257999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.258046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.258057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.258074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.258086 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.360547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.360604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.360619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.360641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.360657 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.463044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.463089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.463145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.463166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.463180 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.566099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.566169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.566180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.566198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.566212 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.669348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.669392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.669405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.669423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.669437 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.772028 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.772065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.772074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.772089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.772098 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.875542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.875869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.875887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.875913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.875933 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.979313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.979382 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.979405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.979434 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:48 crc kubenswrapper[4945]: I1008 15:13:48.979457 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:48Z","lastTransitionTime":"2025-10-08T15:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.082482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.082526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.082537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.082556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.082570 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.184659 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.184709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.184725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.184745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.184759 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.261317 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerStarted","Data":"6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.275775 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.287496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.287539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.287551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.287572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.287588 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.291601 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.325045 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.342185 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.364646 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.381397 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.390468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.390505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.390516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.390534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.390550 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.406031 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.431971 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.449848 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.463551 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.480953 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.493510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.493543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.493553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.493567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.493577 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.498322 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.516174 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.530911 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.546360 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:49Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.595421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.595485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.595508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.595539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.595564 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.681261 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.681431 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.681586 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.681681 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:14:05.681655139 +0000 UTC m=+55.035570200 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.681750 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:05.681738071 +0000 UTC m=+55.035653202 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.700277 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.700341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.700361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.700385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.700409 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.782938 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.783033 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783100 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.783168 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783221 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:05.783196487 +0000 UTC m=+55.137111398 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783357 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783426 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783447 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783521 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:05.783494185 +0000 UTC m=+55.137409106 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783370 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783553 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783563 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:49 crc kubenswrapper[4945]: E1008 15:13:49.783596 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:05.783586537 +0000 UTC m=+55.137501448 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.802920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.802970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.802980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.802999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.803012 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.905077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.905431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.905444 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.905463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:49 crc kubenswrapper[4945]: I1008 15:13:49.905473 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:49Z","lastTransitionTime":"2025-10-08T15:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.007271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.007420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.007496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.007567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.007645 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.023719 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:50 crc kubenswrapper[4945]: E1008 15:13:50.023896 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.023952 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.024065 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:50 crc kubenswrapper[4945]: E1008 15:13:50.024185 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:50 crc kubenswrapper[4945]: E1008 15:13:50.024228 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.111330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.111400 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.111419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.111446 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.111467 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.214418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.214470 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.214491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.214519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.214541 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.270540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.302324 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.318390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.318440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.318480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.318501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.318514 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.322497 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.340044 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.362892 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.381945 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.400576 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.421524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.421584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.421607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.421637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.421659 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.452766 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.479136 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.494150 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.509821 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.522227 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.523627 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.523699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.523722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.523745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.523765 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.535382 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.546802 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.563290 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.585785 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:50Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.626769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.626808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.626819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.626836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.626848 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.729727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.729770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.729779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.729794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.729804 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.833313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.833365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.833381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.833404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.833424 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.936461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.936500 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.936510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.936525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:50 crc kubenswrapper[4945]: I1008 15:13:50.936537 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:50Z","lastTransitionTime":"2025-10-08T15:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.010759 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.010847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.010874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.010906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.010932 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: E1008 15:13:51.032702 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.036634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.036691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.036702 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.036722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.036735 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: E1008 15:13:51.050668 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.058947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.059014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.059034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.059065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.059085 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: E1008 15:13:51.074513 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.078992 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.079059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.079080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.079140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.079159 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: E1008 15:13:51.095045 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.099354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.099398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.099413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.099436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.099453 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: E1008 15:13:51.115055 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: E1008 15:13:51.115234 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.116955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.116988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.116999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.117039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.117054 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.220276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.220334 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.220344 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.220361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.220395 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.295895 4945 generic.go:334] "Generic (PLEG): container finished" podID="e93c84e1-87ae-4485-8ea1-8cd4ac1554a4" containerID="6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc" exitCode=0 Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.296052 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.296038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerDied","Data":"6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.297059 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.297161 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.323318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.323371 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.323384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.323401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.323861 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.334307 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.341608 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.346984 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.349521 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.363451 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.392754 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.407784 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.420767 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.427383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.427438 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.427453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.427474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.427488 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.433412 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.449871 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.464547 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.481298 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.491044 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.504147 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.516719 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.530243 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.531383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.531435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.531445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.531460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.531488 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.549154 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.567800 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.583499 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.613206 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.629538 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.633367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.633406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.633417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.633436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.633447 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.643425 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.659144 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.678984 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.693562 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.707237 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.718634 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.732751 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.735368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.735413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.735423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.735442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.735453 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.744545 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.767997 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.783059 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.793809 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:51Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.837786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.837821 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.837830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.837849 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.837860 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.940422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.940494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.940517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.940548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:51 crc kubenswrapper[4945]: I1008 15:13:51.940570 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:51Z","lastTransitionTime":"2025-10-08T15:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.024141 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:52 crc kubenswrapper[4945]: E1008 15:13:52.024344 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.024183 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:52 crc kubenswrapper[4945]: E1008 15:13:52.024447 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.024160 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:52 crc kubenswrapper[4945]: E1008 15:13:52.024512 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.040378 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.043015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.043131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.043143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.043162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.043175 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.053951 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.072788 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.101335 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.121991 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.142466 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.145283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.145344 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.145362 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.145388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.145431 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.158479 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.174752 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.188856 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.202592 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.218032 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.239261 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.248032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.248100 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.248140 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.248166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.248186 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.255204 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.269312 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.286721 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.303047 4945 generic.go:334] "Generic (PLEG): container finished" podID="e93c84e1-87ae-4485-8ea1-8cd4ac1554a4" containerID="4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3" exitCode=0 Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.303174 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerDied","Data":"4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.303299 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.323918 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.342148 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.351537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.351582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.351595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.351645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.351661 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.360648 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.386329 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.407774 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.420498 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.432821 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.446980 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.454443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.454469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.454477 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.454490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.454499 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.461522 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.475205 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.489935 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.501657 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.513843 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.526082 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.542360 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:52Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.557086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.557136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.557146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.557159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.557168 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.673854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.673891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.673899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.673913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.673924 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.776858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.777147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.777163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.777182 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.777195 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.879930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.879976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.879987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.880005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.880017 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.982396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.982454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.982473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.982497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:52 crc kubenswrapper[4945]: I1008 15:13:52.982516 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:52Z","lastTransitionTime":"2025-10-08T15:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.085645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.085730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.085756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.085793 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.085821 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.122560 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn"] Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.123410 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.126473 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.128018 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.142019 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.158708 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.171392 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.178370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/55656f9e-8162-446e-8b48-5947f35613f6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.178427 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/55656f9e-8162-446e-8b48-5947f35613f6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.178450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv5l5\" (UniqueName: \"kubernetes.io/projected/55656f9e-8162-446e-8b48-5947f35613f6-kube-api-access-kv5l5\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.178478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/55656f9e-8162-446e-8b48-5947f35613f6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.184578 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.188675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.188740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.188761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.188790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.188817 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.194847 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.206922 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.219552 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.231239 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.242440 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.250505 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.266066 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.279567 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/55656f9e-8162-446e-8b48-5947f35613f6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.279641 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/55656f9e-8162-446e-8b48-5947f35613f6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.279681 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/55656f9e-8162-446e-8b48-5947f35613f6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.279732 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv5l5\" (UniqueName: \"kubernetes.io/projected/55656f9e-8162-446e-8b48-5947f35613f6-kube-api-access-kv5l5\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.281165 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/55656f9e-8162-446e-8b48-5947f35613f6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.281369 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/55656f9e-8162-446e-8b48-5947f35613f6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.285266 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/55656f9e-8162-446e-8b48-5947f35613f6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.285410 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.291966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.292039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.292065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.292098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.292180 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.300402 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv5l5\" (UniqueName: \"kubernetes.io/projected/55656f9e-8162-446e-8b48-5947f35613f6-kube-api-access-kv5l5\") pod \"ovnkube-control-plane-749d76644c-p66bn\" (UID: \"55656f9e-8162-446e-8b48-5947f35613f6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.305987 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.310606 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" event={"ID":"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4","Type":"ContainerStarted","Data":"61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.310695 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.325676 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.337832 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.353726 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.382274 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.394912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.394975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.394992 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.395020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.395039 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.404581 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.424320 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.441049 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.446382 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: W1008 15:13:53.462092 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55656f9e_8162_446e_8b48_5947f35613f6.slice/crio-fe90e85f45c8706ee0f6e7c09ff567815de924559e23a43fa14715d888d915c3 WatchSource:0}: Error finding container fe90e85f45c8706ee0f6e7c09ff567815de924559e23a43fa14715d888d915c3: Status 404 returned error can't find the container with id fe90e85f45c8706ee0f6e7c09ff567815de924559e23a43fa14715d888d915c3 Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.466775 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.497067 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.497841 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.497891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.497906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.497927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.497945 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.510964 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.522305 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.538258 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.553228 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.571525 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.585966 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.595123 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.601529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.601566 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.601575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.601592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.601603 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.616072 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.635608 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.645224 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:53Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.704801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.704877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.704902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.704934 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.704958 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.808468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.808549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.808571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.808602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.808626 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.910562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.910596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.910604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.910617 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:53 crc kubenswrapper[4945]: I1008 15:13:53.910627 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:53Z","lastTransitionTime":"2025-10-08T15:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.012978 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.013009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.013020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.013036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.013048 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.024231 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.024365 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.024279 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:54 crc kubenswrapper[4945]: E1008 15:13:54.024519 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:54 crc kubenswrapper[4945]: E1008 15:13:54.024669 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:54 crc kubenswrapper[4945]: E1008 15:13:54.025690 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.115948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.115986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.115997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.116013 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.116028 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.219791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.219846 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.219864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.219889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.219906 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.314393 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" event={"ID":"55656f9e-8162-446e-8b48-5947f35613f6","Type":"ContainerStarted","Data":"fe90e85f45c8706ee0f6e7c09ff567815de924559e23a43fa14715d888d915c3"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.322411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.322466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.322484 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.322505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.322522 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.424726 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.424777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.424789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.424808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.424822 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.527461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.527789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.527799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.527815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.527825 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.612593 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-zj7s7"] Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.613092 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:54 crc kubenswrapper[4945]: E1008 15:13:54.613189 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.630282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.630310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.630322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.630339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.630350 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.650660 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.668522 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.686949 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.695155 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpk49\" (UniqueName: \"kubernetes.io/projected/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-kube-api-access-dpk49\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.695211 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.708936 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.721709 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.732563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.732707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.732723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.732744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.732757 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.742983 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.758179 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.770283 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.783886 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.795698 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.795781 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpk49\" (UniqueName: \"kubernetes.io/projected/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-kube-api-access-dpk49\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.795828 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:54 crc kubenswrapper[4945]: E1008 15:13:54.795935 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:54 crc kubenswrapper[4945]: E1008 15:13:54.795988 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:55.295972571 +0000 UTC m=+44.649887472 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.867799 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.868567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.868612 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.868625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.868644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.868657 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.882042 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.896885 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.909886 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.924944 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.937462 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.948096 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:54Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.970815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.970863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.970878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.970897 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:54 crc kubenswrapper[4945]: I1008 15:13:54.970909 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:54Z","lastTransitionTime":"2025-10-08T15:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.040608 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpk49\" (UniqueName: \"kubernetes.io/projected/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-kube-api-access-dpk49\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.073895 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.073941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.073958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.073979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.073996 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.177424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.177516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.177538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.177573 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.177605 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.280276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.280326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.280342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.280363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.280379 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.301226 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.301423 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.301491 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:56.301471623 +0000 UTC m=+45.655386534 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.319801 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" event={"ID":"55656f9e-8162-446e-8b48-5947f35613f6","Type":"ContainerStarted","Data":"53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.382734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.382802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.382820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.382845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.382863 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.485533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.485644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.485671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.485705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.485727 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.589419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.589476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.589493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.589516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.589581 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.692643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.692703 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.692723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.692747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.692763 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.796761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.796814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.796830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.796850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.796866 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.892825 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.893068 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.893406 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.894197 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.894563 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.894614 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.894981 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.895252 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.895630 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 08 15:13:55 crc kubenswrapper[4945]: E1008 15:13:55.895669 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.899593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.899642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.899652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.899670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:55 crc kubenswrapper[4945]: I1008 15:13:55.899682 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:55Z","lastTransitionTime":"2025-10-08T15:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.002895 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.002979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.003018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.003049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.003070 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.023677 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.023713 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.023738 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:56 crc kubenswrapper[4945]: E1008 15:13:56.023856 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.023925 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:56 crc kubenswrapper[4945]: E1008 15:13:56.024004 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:56 crc kubenswrapper[4945]: E1008 15:13:56.025056 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:56 crc kubenswrapper[4945]: E1008 15:13:56.025296 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.106363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.106418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.106458 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.106494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.106543 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.210603 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.210662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.210674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.210694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.210709 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.311861 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:56 crc kubenswrapper[4945]: E1008 15:13:56.311991 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:56 crc kubenswrapper[4945]: E1008 15:13:56.312043 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:13:58.312025326 +0000 UTC m=+47.665940237 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.313039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.313070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.313081 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.313097 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.313135 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.325553 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/0.log" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.328334 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" exitCode=1 Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.328410 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.329224 4945 scope.go:117] "RemoveContainer" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.333436 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" event={"ID":"55656f9e-8162-446e-8b48-5947f35613f6","Type":"ContainerStarted","Data":"eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.348471 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.358378 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.376711 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.390799 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.401639 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.412265 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.415265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.415309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.415327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.415351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.415367 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.426231 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.441233 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.454260 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.469510 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.480017 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.492961 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.517311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.517355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.517365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.517379 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.517388 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.525196 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:55Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 15:13:54.799817 6244 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 15:13:54.799908 6244 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 15:13:54.799977 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 15:13:54.800055 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 15:13:54.800100 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 15:13:54.800225 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 15:13:54.800294 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 15:13:54.800355 6244 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 15:13:54.799978 6244 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 15:13:54.800484 6244 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 15:13:54.800370 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 15:13:54.800542 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 15:13:54.800392 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 15:13:54.800448 6244 factory.go:656] Stopping watch factory\\\\nI1008 15:13:54.800639 6244 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.541561 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.556324 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.568646 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.585997 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.602429 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.614858 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.619058 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.619173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.619193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.619215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.619231 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.633012 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.647715 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.680949 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.701992 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.718222 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.724282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.724357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.724384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.724435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.724463 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.738147 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.754268 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.770253 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.784380 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.797453 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.810217 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.827071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.827192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.827217 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.827247 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.827272 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.831467 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.864647 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:55Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 15:13:54.799817 6244 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 15:13:54.799908 6244 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 15:13:54.799977 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 15:13:54.800055 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 15:13:54.800100 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 15:13:54.800225 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 15:13:54.800294 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 15:13:54.800355 6244 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 15:13:54.799978 6244 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 15:13:54.800484 6244 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 15:13:54.800370 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 15:13:54.800542 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 15:13:54.800392 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 15:13:54.800448 6244 factory.go:656] Stopping watch factory\\\\nI1008 15:13:54.800639 6244 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.881921 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.896146 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:56Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.929694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.929757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.929776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.929798 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:56 crc kubenswrapper[4945]: I1008 15:13:56.929813 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:56Z","lastTransitionTime":"2025-10-08T15:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.024889 4945 scope.go:117] "RemoveContainer" containerID="9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.032263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.032331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.032347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.032403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.032422 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.134683 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.134723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.134732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.134746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.134754 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.237977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.238012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.238020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.238033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.238043 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.337646 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/0.log" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.339202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.339241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.339253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.339270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.339283 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.340431 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.340848 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.342420 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.343869 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.344090 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.353400 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.365633 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.376835 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.394076 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.408810 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.419739 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.432023 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.442295 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.442352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.442369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.442757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.442812 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.447220 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.457779 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.469519 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.484166 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.502446 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.517100 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.538209 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:55Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 15:13:54.799817 6244 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 15:13:54.799908 6244 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 15:13:54.799977 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 15:13:54.800055 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 15:13:54.800100 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 15:13:54.800225 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 15:13:54.800294 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 15:13:54.800355 6244 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 15:13:54.799978 6244 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 15:13:54.800484 6244 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 15:13:54.800370 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 15:13:54.800542 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 15:13:54.800392 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 15:13:54.800448 6244 factory.go:656] Stopping watch factory\\\\nI1008 15:13:54.800639 6244 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.545939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.546057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.546077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.546133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.546153 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.554201 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.564058 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.571760 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.581878 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.590026 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.600480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.617572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:55Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 15:13:54.799817 6244 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 15:13:54.799908 6244 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 15:13:54.799977 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 15:13:54.800055 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 15:13:54.800100 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 15:13:54.800225 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 15:13:54.800294 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 15:13:54.800355 6244 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 15:13:54.799978 6244 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 15:13:54.800484 6244 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 15:13:54.800370 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 15:13:54.800542 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 15:13:54.800392 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 15:13:54.800448 6244 factory.go:656] Stopping watch factory\\\\nI1008 15:13:54.800639 6244 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.628833 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.647269 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.649081 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.649149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.649162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.649180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.649193 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.663055 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.673506 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.688480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.700083 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.711952 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.732963 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.747392 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.751647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.751684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.751696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.751716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.751732 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.759037 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.772249 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.787452 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.806842 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.853830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.853864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.853873 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.853885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.853895 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.956620 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.956678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.956687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.956708 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:57 crc kubenswrapper[4945]: I1008 15:13:57.956720 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:57Z","lastTransitionTime":"2025-10-08T15:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.023586 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.023705 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:13:58 crc kubenswrapper[4945]: E1008 15:13:58.023768 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.023791 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.023832 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:58 crc kubenswrapper[4945]: E1008 15:13:58.023942 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:13:58 crc kubenswrapper[4945]: E1008 15:13:58.024099 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:13:58 crc kubenswrapper[4945]: E1008 15:13:58.024314 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.059795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.059920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.059932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.059945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.059955 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.162581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.162652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.162672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.162700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.162723 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.265198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.265289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.265311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.265340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.265357 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.332635 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:13:58 crc kubenswrapper[4945]: E1008 15:13:58.332834 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:58 crc kubenswrapper[4945]: E1008 15:13:58.332921 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:02.332900474 +0000 UTC m=+51.686815375 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.348861 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/1.log" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.349590 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/0.log" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.352301 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309" exitCode=1 Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.352384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.352471 4945 scope.go:117] "RemoveContainer" containerID="c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.353399 4945 scope.go:117] "RemoveContainer" containerID="b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309" Oct 08 15:13:58 crc kubenswrapper[4945]: E1008 15:13:58.353586 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.367722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.367760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.367769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.367783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.367793 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.372709 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.386000 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.403489 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.421033 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.431166 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.446812 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.464400 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.470883 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.470925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.470935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.470952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.470966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.478865 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.500480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.514580 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.527063 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.538642 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.551034 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.562712 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.573414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.573482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.573497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.573515 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.573529 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.578285 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.594979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c47c06f5ed31e612f9ad6526fde109b11f2d40279ada8c27619a96634cd0c21d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:55Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 15:13:54.799817 6244 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 15:13:54.799908 6244 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 15:13:54.799977 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 15:13:54.800055 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 15:13:54.800100 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 15:13:54.800225 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 15:13:54.800294 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 15:13:54.800355 6244 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 15:13:54.799978 6244 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 15:13:54.800484 6244 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 15:13:54.800370 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 15:13:54.800542 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 15:13:54.800392 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 15:13:54.800448 6244 factory.go:656] Stopping watch factory\\\\nI1008 15:13:54.800639 6244 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.606702 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:58Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.676212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.676280 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.676304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.676333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.676356 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.779163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.779215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.779228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.779245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.779256 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.882205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.882246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.882257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.882273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.882284 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.984962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.985018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.985034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.985054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:58 crc kubenswrapper[4945]: I1008 15:13:58.985067 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:58Z","lastTransitionTime":"2025-10-08T15:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.087758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.087821 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.087839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.087861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.087878 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.190264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.190334 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.190351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.190373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.190390 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.292540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.292576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.292586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.292622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.292632 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.357103 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/1.log" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.361750 4945 scope.go:117] "RemoveContainer" containerID="b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309" Oct 08 15:13:59 crc kubenswrapper[4945]: E1008 15:13:59.361999 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.380340 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.392979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.395201 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.395265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.395284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.395311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.395331 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.409269 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.435653 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.472882 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.491491 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.497718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.497766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.497784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.497808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.497824 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.503370 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.513015 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.524330 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.534081 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.544135 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.551247 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.560514 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.572763 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.587462 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.599933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.599963 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.599972 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.599986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.599996 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.604235 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.616960 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:59Z is after 2025-08-24T17:21:41Z" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.702257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.702321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.702339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.702364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.702381 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.804784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.804858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.804881 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.804913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.804938 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.907776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.907842 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.907867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.907896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:13:59 crc kubenswrapper[4945]: I1008 15:13:59.907922 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:13:59Z","lastTransitionTime":"2025-10-08T15:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.011394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.011469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.011493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.011523 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.011546 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.023719 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.023817 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:00 crc kubenswrapper[4945]: E1008 15:14:00.023972 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.024013 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:00 crc kubenswrapper[4945]: E1008 15:14:00.024191 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:00 crc kubenswrapper[4945]: E1008 15:14:00.024283 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.024401 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:00 crc kubenswrapper[4945]: E1008 15:14:00.024583 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.113105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.113153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.113161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.113175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.113184 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.215961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.216006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.216018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.216036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.216048 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.319514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.319592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.319615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.319645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.319666 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.422543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.422618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.422641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.422673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.422692 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.525163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.525219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.525235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.525256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.525271 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.627551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.627638 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.627654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.627673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.627690 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.730671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.730729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.730744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.730771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.730788 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.832826 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.832879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.832891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.832909 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.832923 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.935090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.935137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.935146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.935158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:00 crc kubenswrapper[4945]: I1008 15:14:00.935167 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:00Z","lastTransitionTime":"2025-10-08T15:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.037938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.037988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.038005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.038024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.038035 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.140255 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.140304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.140314 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.140327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.140337 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.243544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.243599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.243610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.243629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.243641 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.346215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.346258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.346268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.346294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.346303 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.449790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.449841 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.449855 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.449873 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.449886 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.461080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.461190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.461215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.461246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.461273 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: E1008 15:14:01.484360 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:01Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.488912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.488954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.488966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.488983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.488995 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: E1008 15:14:01.503574 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:01Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.507401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.507469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.507487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.507875 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.508232 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: E1008 15:14:01.523950 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:01Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.527879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.527946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.527974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.528005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.528029 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: E1008 15:14:01.545096 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:01Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.549510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.549533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.549541 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.549553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.549562 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: E1008 15:14:01.567091 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:01Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:01 crc kubenswrapper[4945]: E1008 15:14:01.567355 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.568786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.568822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.568833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.568859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.568880 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.670781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.670839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.670861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.670884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.670898 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.773639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.773690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.773705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.773725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.773739 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.876746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.876787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.876798 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.876813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.876822 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.979861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.979941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.979976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.980006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:01 crc kubenswrapper[4945]: I1008 15:14:01.980026 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:01Z","lastTransitionTime":"2025-10-08T15:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.023829 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.023862 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.023950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:02 crc kubenswrapper[4945]: E1008 15:14:02.024063 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.024150 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:02 crc kubenswrapper[4945]: E1008 15:14:02.024413 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:02 crc kubenswrapper[4945]: E1008 15:14:02.024591 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:02 crc kubenswrapper[4945]: E1008 15:14:02.024778 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.044968 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.059532 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.073557 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.082691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.082719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.082727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.082740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.082749 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.086774 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.100937 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.114277 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.126323 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.139169 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.148708 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.159473 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.175531 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.185012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.185078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.185091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.185123 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.185134 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.185005 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.201493 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.214314 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.227893 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.249607 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.263692 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:02Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.287416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.287449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.287461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.287478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.287490 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.371551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:02 crc kubenswrapper[4945]: E1008 15:14:02.371675 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:14:02 crc kubenswrapper[4945]: E1008 15:14:02.371742 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:10.371724205 +0000 UTC m=+59.725639116 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.389776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.389819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.389832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.389850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.389862 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.492457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.492498 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.492507 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.492523 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.492533 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.595288 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.595329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.595343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.595358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.595367 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.698740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.698791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.698809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.698831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.698848 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.802009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.802076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.802098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.802167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.802191 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.905631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.905690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.905707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.905730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.905751 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:02Z","lastTransitionTime":"2025-10-08T15:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:02 crc kubenswrapper[4945]: I1008 15:14:02.994239 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.005672 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.008553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.008595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.008605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.008626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.008639 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.015330 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.037668 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.057786 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.077022 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.094393 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.108347 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.111625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.111660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.111672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.111689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.111704 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.124556 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.136027 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.150619 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.161602 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.175773 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.189001 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.201418 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.214258 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.214626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.214706 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.214734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.214760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.214779 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.226226 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.243433 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.264095 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:03Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.317886 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.317940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.317958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.317983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.318003 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.421211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.421290 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.421344 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.421376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.421399 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.524244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.524306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.524322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.524346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.524363 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.627856 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.627923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.627941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.627968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.627987 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.731454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.731624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.731647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.731673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.731691 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.834839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.835155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.835172 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.835194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.835208 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.939867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.939979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.939999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.940025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:03 crc kubenswrapper[4945]: I1008 15:14:03.940042 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:03Z","lastTransitionTime":"2025-10-08T15:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.024281 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.024333 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.024333 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.024428 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:04 crc kubenswrapper[4945]: E1008 15:14:04.024631 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:04 crc kubenswrapper[4945]: E1008 15:14:04.024806 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:04 crc kubenswrapper[4945]: E1008 15:14:04.024930 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:04 crc kubenswrapper[4945]: E1008 15:14:04.025018 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.042229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.042280 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.042298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.042319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.042336 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.145498 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.145666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.145681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.145697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.145708 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.249278 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.249337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.249374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.249402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.249424 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.352837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.352885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.352902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.352925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.352943 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.456269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.456368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.456401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.456610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.456637 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.560096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.560198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.560222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.560252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.560274 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.662290 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.662327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.662338 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.662353 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.662368 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.765166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.765253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.765284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.765331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.765356 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.868059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.868102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.868133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.868153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.868167 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.971351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.971384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.971392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.971405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:04 crc kubenswrapper[4945]: I1008 15:14:04.971414 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:04Z","lastTransitionTime":"2025-10-08T15:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.074143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.074198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.074207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.074221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.074231 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.176710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.176746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.176754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.176783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.176798 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.279895 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.279973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.279991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.280015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.280032 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.383266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.383337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.383355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.383380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.383398 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.486733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.486787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.486804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.486827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.486845 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.589465 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.589535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.589560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.589590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.589613 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.692363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.692406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.692420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.692440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.692454 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.705476 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.705729 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:14:37.705693547 +0000 UTC m=+87.059608448 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.705801 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.705948 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.706070 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:37.706039406 +0000 UTC m=+87.059954447 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.800179 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.800241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.800263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.800292 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.800315 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.806560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.806629 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.806691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.806887 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.806895 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.806923 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.806948 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.806945 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.806987 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.807010 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.807011 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:37.806976688 +0000 UTC m=+87.160891629 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.807097 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:37.807075121 +0000 UTC m=+87.160990062 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:05 crc kubenswrapper[4945]: E1008 15:14:05.807175 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:37.807154853 +0000 UTC m=+87.161069804 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.903152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.903224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.903258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.903287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:05 crc kubenswrapper[4945]: I1008 15:14:05.903307 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:05Z","lastTransitionTime":"2025-10-08T15:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.006515 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.006599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.006624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.006652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.006673 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.023846 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.023890 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.023882 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.023981 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:06 crc kubenswrapper[4945]: E1008 15:14:06.024088 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:06 crc kubenswrapper[4945]: E1008 15:14:06.024219 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:06 crc kubenswrapper[4945]: E1008 15:14:06.024650 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:06 crc kubenswrapper[4945]: E1008 15:14:06.024507 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.109385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.109451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.109462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.109478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.109489 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.211735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.211813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.211837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.211869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.211890 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.313879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.313944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.313955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.313969 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.313981 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.417208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.417346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.417373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.417408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.417433 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.519937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.520012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.520039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.520200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.520259 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.622479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.622535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.622559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.622585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.622605 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.726340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.726693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.727208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.727385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.727524 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.830097 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.830399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.830525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.830687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.830810 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.934064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.934212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.934239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.934271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:06 crc kubenswrapper[4945]: I1008 15:14:06.934297 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:06Z","lastTransitionTime":"2025-10-08T15:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.037225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.037299 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.037324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.037354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.037377 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.140800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.140848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.140864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.140882 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.140895 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.243566 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.243628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.243642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.243664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.243678 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.346137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.346203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.346227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.346259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.346282 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.448206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.448260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.448272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.448290 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.448303 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.551874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.551930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.551946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.551966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.551984 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.654304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.654352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.654364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.654383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.654397 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.757467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.757535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.757549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.757567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.757578 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.860091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.860184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.860204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.860229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.860247 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.962271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.962303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.962310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.962323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:07 crc kubenswrapper[4945]: I1008 15:14:07.962332 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:07Z","lastTransitionTime":"2025-10-08T15:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.023479 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.023529 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.023571 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:08 crc kubenswrapper[4945]: E1008 15:14:08.023728 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.023755 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:08 crc kubenswrapper[4945]: E1008 15:14:08.023850 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:08 crc kubenswrapper[4945]: E1008 15:14:08.023991 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:08 crc kubenswrapper[4945]: E1008 15:14:08.023916 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.066631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.066703 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.066729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.066774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.066798 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.169891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.169955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.169975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.169998 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.170016 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.272796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.273456 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.273508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.273540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.273565 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.377248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.377348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.377373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.377396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.377414 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.480146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.480203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.480218 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.480235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.480248 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.583894 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.583955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.583966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.583984 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.583999 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.687717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.687849 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.687876 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.687905 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.687924 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.790817 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.790891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.790910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.790938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.790956 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.893646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.894354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.894386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.894415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.894439 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.997501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.997551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.997568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.997585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:08 crc kubenswrapper[4945]: I1008 15:14:08.997597 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:08Z","lastTransitionTime":"2025-10-08T15:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.100089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.100165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.100178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.100193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.100204 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.202175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.202218 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.202230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.202245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.202255 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.305597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.305694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.305730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.305762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.305799 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.408577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.408635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.408650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.408672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.408687 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.511096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.511147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.511156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.511169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.511177 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.613234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.613283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.613294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.613313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.613327 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.716126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.716196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.716207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.716239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.716253 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.818936 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.819001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.819026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.819052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.819090 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.922388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.922459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.922478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.922504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:09 crc kubenswrapper[4945]: I1008 15:14:09.922517 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:09Z","lastTransitionTime":"2025-10-08T15:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.023832 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.023947 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:10 crc kubenswrapper[4945]: E1008 15:14:10.024043 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.024100 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.024922 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:10 crc kubenswrapper[4945]: E1008 15:14:10.025067 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:10 crc kubenswrapper[4945]: E1008 15:14:10.025206 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:10 crc kubenswrapper[4945]: E1008 15:14:10.025373 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.026740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.026814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.026838 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.026866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.026887 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.129229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.129283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.129368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.129396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.129423 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.231705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.231750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.231775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.231790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.231800 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.335024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.335065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.335074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.335094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.335129 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.438233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.438286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.438297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.438317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.438329 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.456889 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:10 crc kubenswrapper[4945]: E1008 15:14:10.457063 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:14:10 crc kubenswrapper[4945]: E1008 15:14:10.457165 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:26.457144411 +0000 UTC m=+75.811059322 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.540659 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.540747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.540806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.540836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.540858 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.642769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.642822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.642834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.642849 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.642860 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.745143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.745187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.745199 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.745215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.745227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.847386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.847433 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.847443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.847458 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.847467 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.949810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.949851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.949859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.949874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:10 crc kubenswrapper[4945]: I1008 15:14:10.949889 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:10Z","lastTransitionTime":"2025-10-08T15:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.052513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.052568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.052583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.052604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.052619 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.155856 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.155954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.155987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.156045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.156074 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.259401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.259443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.259455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.259471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.259482 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.361887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.361932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.361940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.361956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.361967 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.465236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.465348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.465362 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.465388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.465403 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.567935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.567986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.568000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.568022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.568036 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.671413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.671550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.671574 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.671602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.671620 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.746725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.746788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.746810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.746843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.746864 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: E1008 15:14:11.760791 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:11Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.764976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.765010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.765019 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.765038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.765048 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: E1008 15:14:11.783215 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:11Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.788227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.788275 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.788285 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.788301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.788312 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: E1008 15:14:11.802916 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:11Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.807322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.807394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.807408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.807437 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.807451 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: E1008 15:14:11.825100 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:11Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.829597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.829647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.829664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.829685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.829700 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: E1008 15:14:11.848815 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:11Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:11 crc kubenswrapper[4945]: E1008 15:14:11.848964 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.850479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.850519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.850535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.850554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.850572 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.953068 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.953155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.953167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.953191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:11 crc kubenswrapper[4945]: I1008 15:14:11.953203 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:11Z","lastTransitionTime":"2025-10-08T15:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.023379 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:12 crc kubenswrapper[4945]: E1008 15:14:12.023526 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.023844 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:12 crc kubenswrapper[4945]: E1008 15:14:12.023897 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.023944 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:12 crc kubenswrapper[4945]: E1008 15:14:12.023988 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.024186 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:12 crc kubenswrapper[4945]: E1008 15:14:12.024263 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.028102 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.041941 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.052596 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.055484 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.055533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.055547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.055570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.055585 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.076076 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.090339 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.105448 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.119478 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.134785 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.148910 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.158154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.158216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.158229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.158254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.158269 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.163892 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.178197 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.193967 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.210747 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.225591 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.254965 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.260032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.260091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.260133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.260160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.260179 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.269798 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.284535 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.296836 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.313076 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.327672 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.340688 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.355682 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.362613 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.362668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.362679 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.362696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.362708 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.378954 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.389576 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.413145 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.427537 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.441852 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.454961 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.464862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.464912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.464924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.464944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.464957 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.467027 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.479741 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.492290 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.501071 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.511896 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.522444 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.532970 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.545296 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.556088 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:12Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.567788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.567823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.567833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.567848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.567857 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.670497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.670559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.670572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.670586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.670596 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.773242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.773272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.773293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.773305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.773314 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.876374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.876437 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.876449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.876464 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.876476 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.978474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.978518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.978530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.978548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:12 crc kubenswrapper[4945]: I1008 15:14:12.978560 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:12Z","lastTransitionTime":"2025-10-08T15:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.081439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.081486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.081496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.081513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.081523 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.184298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.184352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.184368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.184395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.184423 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.287225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.287259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.287271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.287286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.287297 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.391071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.391504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.391702 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.391897 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.392145 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.495402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.495469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.495483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.495509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.495525 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.598883 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.598930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.598944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.598963 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.598978 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.701995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.702063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.702073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.702089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.702098 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.804889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.805210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.805222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.805241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.805254 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.908799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.908863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.908879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.908902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:13 crc kubenswrapper[4945]: I1008 15:14:13.908939 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:13Z","lastTransitionTime":"2025-10-08T15:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.012265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.012340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.012367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.012398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.012420 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.024273 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.024394 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:14 crc kubenswrapper[4945]: E1008 15:14:14.024397 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.024437 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:14 crc kubenswrapper[4945]: E1008 15:14:14.024485 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.024635 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:14 crc kubenswrapper[4945]: E1008 15:14:14.024625 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:14 crc kubenswrapper[4945]: E1008 15:14:14.025076 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.025326 4945 scope.go:117] "RemoveContainer" containerID="b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.114903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.114933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.114941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.114955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.114964 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.218313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.218376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.218395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.218436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.218454 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.321884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.321957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.321979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.322006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.322023 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.425651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.425711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.425735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.425763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.425784 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.528399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.528448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.528460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.528476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.528511 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.631548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.631598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.631617 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.631641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.631658 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.734941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.735046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.735078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.735101 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.735151 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.838315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.838377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.838394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.838419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.838437 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.941612 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.941650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.941660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.941675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:14 crc kubenswrapper[4945]: I1008 15:14:14.941684 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:14Z","lastTransitionTime":"2025-10-08T15:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.044710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.044939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.044947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.044960 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.044969 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.147692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.147746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.147763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.147783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.147797 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.250622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.250688 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.250712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.250741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.250766 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.353861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.353924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.353938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.353961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.353977 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.415066 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/1.log" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.417559 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.417991 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.437213 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.448514 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.455890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.455917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.455925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.455938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.455948 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.460000 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.473964 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.484894 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.496786 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.509771 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.524767 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.537194 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.549790 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.558272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.558305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.558313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.558328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.558336 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.560915 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.570853 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.579585 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.590653 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.601029 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.611990 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.628595 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.641837 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.660537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.660565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.660574 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.660588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.660596 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.763403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.763449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.763462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.763480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.763492 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.866296 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.866358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.866378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.866404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.866423 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.969045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.969095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.969129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.969154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:15 crc kubenswrapper[4945]: I1008 15:14:15.969167 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:15Z","lastTransitionTime":"2025-10-08T15:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.023418 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.023485 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.023434 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:16 crc kubenswrapper[4945]: E1008 15:14:16.023574 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.023590 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:16 crc kubenswrapper[4945]: E1008 15:14:16.023744 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:16 crc kubenswrapper[4945]: E1008 15:14:16.023851 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:16 crc kubenswrapper[4945]: E1008 15:14:16.023999 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.071578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.071631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.071644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.071694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.071708 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.174165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.174228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.174243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.174261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.174274 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.276779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.276840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.276857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.276880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.276900 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.380102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.380184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.380201 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.380224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.380241 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.424138 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/2.log" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.425469 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/1.log" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.429120 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf" exitCode=1 Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.429158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.429195 4945 scope.go:117] "RemoveContainer" containerID="b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.430047 4945 scope.go:117] "RemoveContainer" containerID="4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf" Oct 08 15:14:16 crc kubenswrapper[4945]: E1008 15:14:16.430282 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.447687 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.462602 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.479024 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.482812 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.482880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.482891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.482914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.482926 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.491799 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.510436 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.520378 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.530293 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.546164 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b189f29055cf71c31810b94811e3a7d4d543230eee323a4530bdaded5bfa6309\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:13:57Z\\\",\\\"message\\\":\\\"n switch crc\\\\nI1008 15:13:57.494603 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494608 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-z7jwc\\\\nI1008 15:13:57.494612 6487 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-z7jwc in node crc\\\\nI1008 15:13:57.494615 6487 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-z7jwc after 0 failed attempt(s)\\\\nF1008 15:13:57.494616 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:13:57Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:13:57.494624 6487 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.555285 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.563775 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.576334 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.585366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.585402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.585413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.585430 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.585441 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.586957 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.599214 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.607751 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.628945 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.641540 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.652698 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.668102 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:16Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.688286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.688343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.688366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.688395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.688417 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.791614 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.791954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.792138 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.792266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.792381 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.894831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.895513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.895604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.895745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.895882 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.998324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.998377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.998388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.998404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:16 crc kubenswrapper[4945]: I1008 15:14:16.998416 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:16Z","lastTransitionTime":"2025-10-08T15:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.101524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.101577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.101588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.101607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.101621 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.204455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.204513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.204532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.204559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.204583 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.307461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.307506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.307517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.307537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.307548 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.410952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.410997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.411009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.411025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.411040 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.433869 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/2.log" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.437658 4945 scope.go:117] "RemoveContainer" containerID="4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf" Oct 08 15:14:17 crc kubenswrapper[4945]: E1008 15:14:17.437815 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.448749 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.460235 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.472308 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.488194 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.497814 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.512595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.512634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.512646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.512666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.512678 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.514063 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.527146 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.540377 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.553215 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.563377 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.576509 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.591534 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.602383 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.613562 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.615032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.615074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.615089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.615134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.615151 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.625950 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.636336 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.646567 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.656829 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:17Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.717804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.717836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.717844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.717857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.717866 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.820189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.820229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.820241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.820258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.820267 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.922993 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.923036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.923049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.923065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:17 crc kubenswrapper[4945]: I1008 15:14:17.923077 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:17Z","lastTransitionTime":"2025-10-08T15:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.024304 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.024313 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.024318 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:18 crc kubenswrapper[4945]: E1008 15:14:18.024530 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:18 crc kubenswrapper[4945]: E1008 15:14:18.024680 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:18 crc kubenswrapper[4945]: E1008 15:14:18.024800 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.025316 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:18 crc kubenswrapper[4945]: E1008 15:14:18.025568 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.025679 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.025701 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.025712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.025728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.025741 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.128256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.128305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.128315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.128332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.128344 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.231242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.231303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.231312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.231328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.231337 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.334762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.334803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.334819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.334864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.334885 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.437297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.437337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.437348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.437362 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.437372 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.540025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.540056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.540064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.540076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.540085 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.642201 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.642242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.642255 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.642272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.642285 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.744356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.744390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.744401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.744416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.744428 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.846482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.846761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.846954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.847135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.847291 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.949942 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.949981 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.949992 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.950007 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:18 crc kubenswrapper[4945]: I1008 15:14:18.950018 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:18Z","lastTransitionTime":"2025-10-08T15:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.053052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.053089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.053103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.053141 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.053156 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.156348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.156392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.156406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.156422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.156433 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.259736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.260189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.260205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.260224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.260238 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.362423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.362513 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.362539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.362567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.362584 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.465394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.465493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.465511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.465536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.465551 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.569305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.569355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.569365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.569384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.569394 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.671971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.672021 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.672030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.672046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.672055 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.774638 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.774680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.774690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.774705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.774714 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.877227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.877273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.877284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.877298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.877309 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.979833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.979893 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.979906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.979927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:19 crc kubenswrapper[4945]: I1008 15:14:19.979941 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:19Z","lastTransitionTime":"2025-10-08T15:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.023926 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.023949 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:20 crc kubenswrapper[4945]: E1008 15:14:20.024749 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.024035 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.023989 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:20 crc kubenswrapper[4945]: E1008 15:14:20.024953 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:20 crc kubenswrapper[4945]: E1008 15:14:20.025590 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:20 crc kubenswrapper[4945]: E1008 15:14:20.025736 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.083200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.083265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.083280 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.083297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.083310 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.185538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.185588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.185600 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.185618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.185630 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.288894 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.288957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.288969 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.288987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.289000 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.391615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.391673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.391690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.391714 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.391731 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.495859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.495904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.495915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.495931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.495942 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.598185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.598231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.598242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.598258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.598268 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.700522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.700579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.700592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.700611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.700623 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.803882 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.803949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.803964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.803990 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.804010 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.907869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.908013 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.908467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.908532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:20 crc kubenswrapper[4945]: I1008 15:14:20.908558 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:20Z","lastTransitionTime":"2025-10-08T15:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.012646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.012696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.012708 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.012726 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.012770 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.115583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.115677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.115727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.115753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.115803 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.218783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.218821 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.218834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.218851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.218864 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.321067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.321100 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.321127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.321142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.321165 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.424523 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.424573 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.424584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.424599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.424608 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.526695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.526743 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.526756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.526775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.526789 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.629029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.629061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.629069 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.629083 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.629091 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.731378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.731428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.731440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.731459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.731475 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.833836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.833881 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.833891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.833903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.833912 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.916258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.916331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.916342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.916360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.916372 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: E1008 15:14:21.927472 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:21Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.931241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.931276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.931287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.931303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.931315 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: E1008 15:14:21.941677 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:21Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.944539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.944580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.944593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.944610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.944622 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: E1008 15:14:21.956280 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:21Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.959433 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.959473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.959481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.959495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.959505 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: E1008 15:14:21.970687 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:21Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.973664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.973694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.973704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.973716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.973725 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:21 crc kubenswrapper[4945]: E1008 15:14:21.983617 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:21Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:21 crc kubenswrapper[4945]: E1008 15:14:21.983725 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.985149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.985179 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.985189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.985204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:21 crc kubenswrapper[4945]: I1008 15:14:21.985216 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:21Z","lastTransitionTime":"2025-10-08T15:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.023632 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:22 crc kubenswrapper[4945]: E1008 15:14:22.024022 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.027573 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.027633 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.028058 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:22 crc kubenswrapper[4945]: E1008 15:14:22.028642 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:22 crc kubenswrapper[4945]: E1008 15:14:22.028660 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:22 crc kubenswrapper[4945]: E1008 15:14:22.028755 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.039638 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.053662 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.068769 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.079392 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.087256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.087302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.087315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.087335 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.087349 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.100208 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.113691 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.123770 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.135265 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.148473 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.160156 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.170800 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.180232 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.189802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.189835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.189844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.189858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.189867 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.192691 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.201875 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.213183 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.241138 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.272735 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.292339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.292387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.292398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.292418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.292429 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.294550 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:22Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.394827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.394869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.394878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.394890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.394899 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.497941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.497987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.498005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.498028 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.498044 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.600683 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.600730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.600748 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.600770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.600790 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.702374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.702429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.702441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.702461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.702475 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.806401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.806457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.806471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.806490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.806502 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.909695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.909732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.909743 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.909760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:22 crc kubenswrapper[4945]: I1008 15:14:22.909771 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:22Z","lastTransitionTime":"2025-10-08T15:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.012208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.012241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.012250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.012263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.012272 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.114973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.115034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.115043 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.115056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.115065 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.217224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.217262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.217271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.217305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.217316 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.319832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.319872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.319882 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.319899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.319911 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.421721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.421769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.421782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.421797 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.421809 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.524224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.524284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.524298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.524317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.524329 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.627033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.627136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.627150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.627168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.627182 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.730486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.730540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.730551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.730570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.730582 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.834396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.834468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.834496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.834530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.834554 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.936937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.937216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.937382 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.937479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:23 crc kubenswrapper[4945]: I1008 15:14:23.937567 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:23Z","lastTransitionTime":"2025-10-08T15:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.023417 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.023493 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.023565 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.023592 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:24 crc kubenswrapper[4945]: E1008 15:14:24.024368 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:24 crc kubenswrapper[4945]: E1008 15:14:24.024472 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:24 crc kubenswrapper[4945]: E1008 15:14:24.024516 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:24 crc kubenswrapper[4945]: E1008 15:14:24.024554 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.039867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.039925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.039938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.039956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.039968 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.142030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.142079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.142096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.142151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.142169 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.244666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.244724 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.244742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.244766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.244785 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.347744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.347795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.347811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.347831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.347848 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.450098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.450176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.450187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.450204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.450215 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.552979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.553021 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.553032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.553049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.553061 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.661332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.661390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.661402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.661419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.661473 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.763795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.763847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.763862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.763882 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.763898 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.867117 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.867188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.867201 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.867220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.867234 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.971067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.971147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.971161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.971186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:24 crc kubenswrapper[4945]: I1008 15:14:24.971198 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:24Z","lastTransitionTime":"2025-10-08T15:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.074604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.074765 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.074787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.074809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.074823 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.176766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.176813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.176825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.176845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.176858 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.280026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.280070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.280086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.280137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.280156 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.383026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.383146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.383163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.383181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.383193 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.486581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.486656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.486680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.486711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.486734 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.603223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.603264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.603275 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.603291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.603303 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.706086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.706220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.706236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.706254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.706308 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.808209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.808289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.808302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.808321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.808334 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.910441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.910481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.910492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.910508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:25 crc kubenswrapper[4945]: I1008 15:14:25.910521 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:25Z","lastTransitionTime":"2025-10-08T15:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.013135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.013199 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.013208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.013221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.013230 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.024346 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.024392 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:26 crc kubenswrapper[4945]: E1008 15:14:26.024504 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.024539 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:26 crc kubenswrapper[4945]: E1008 15:14:26.024652 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:26 crc kubenswrapper[4945]: E1008 15:14:26.024749 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.024831 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:26 crc kubenswrapper[4945]: E1008 15:14:26.024915 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.115874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.115919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.115929 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.115945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.115955 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.218073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.218129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.218142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.218156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.218168 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.320875 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.320913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.320924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.320941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.320953 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.424349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.424392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.424402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.424418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.424431 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.527215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.527252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.527261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.527282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.527293 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.536077 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:26 crc kubenswrapper[4945]: E1008 15:14:26.536309 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:14:26 crc kubenswrapper[4945]: E1008 15:14:26.536388 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:14:58.536364643 +0000 UTC m=+107.890279534 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.629881 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.629920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.629931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.629951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.629964 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.733270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.733330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.733346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.733366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.733385 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.836022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.836087 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.836104 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.836159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.836181 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.938238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.938288 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.938301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.938320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:26 crc kubenswrapper[4945]: I1008 15:14:26.938334 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:26Z","lastTransitionTime":"2025-10-08T15:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.041813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.041939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.041956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.041977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.042020 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.144461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.144504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.144516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.144531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.144543 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.247167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.247215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.247227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.247246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.247258 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.349682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.349719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.349730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.349746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.349758 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.451584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.451646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.451659 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.451678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.451691 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.558381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.558435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.558448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.558466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.558479 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.661684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.661728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.661741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.661757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.661768 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.764162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.764202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.764210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.764225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.764235 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.867525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.867568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.867577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.867591 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.867604 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.970026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.970076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.970088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.970105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:27 crc kubenswrapper[4945]: I1008 15:14:27.970142 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:27Z","lastTransitionTime":"2025-10-08T15:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.023316 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.023396 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.023343 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.023383 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:28 crc kubenswrapper[4945]: E1008 15:14:28.023530 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:28 crc kubenswrapper[4945]: E1008 15:14:28.023733 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:28 crc kubenswrapper[4945]: E1008 15:14:28.023834 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:28 crc kubenswrapper[4945]: E1008 15:14:28.023909 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.072994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.073037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.073049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.073068 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.073081 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.175418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.175471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.175487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.175506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.175519 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.278909 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.278982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.279008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.279036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.279058 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.382090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.382186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.382206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.382232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.382253 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.485202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.485375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.485406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.485440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.485467 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.589253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.589321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.589342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.589368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.589391 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.691906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.691957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.691976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.691997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.692012 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.794381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.794426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.794436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.794453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.794466 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.896834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.896920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.896934 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.896954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:28 crc kubenswrapper[4945]: I1008 15:14:28.896966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:28Z","lastTransitionTime":"2025-10-08T15:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.000270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.000323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.000334 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.000352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.000364 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.102785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.103246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.103449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.103611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.103762 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.206226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.206779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.206970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.207204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.207381 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.311387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.311443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.311459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.311481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.311497 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.418809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.418871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.418897 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.418923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.418942 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.521992 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.522074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.522092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.522184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.522211 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.624908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.625004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.625024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.625045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.625060 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.728222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.728260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.728268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.728282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.728291 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.830685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.830763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.830784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.830834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.830851 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.933634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.933681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.933692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.933710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:29 crc kubenswrapper[4945]: I1008 15:14:29.933721 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:29Z","lastTransitionTime":"2025-10-08T15:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.023557 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.023622 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.023558 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:30 crc kubenswrapper[4945]: E1008 15:14:30.023735 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.023807 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:30 crc kubenswrapper[4945]: E1008 15:14:30.023908 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:30 crc kubenswrapper[4945]: E1008 15:14:30.024079 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:30 crc kubenswrapper[4945]: E1008 15:14:30.024188 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.035787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.035835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.035852 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.035876 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.035894 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.138497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.138571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.138594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.138632 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.138692 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.241770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.241807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.241818 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.241833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.241843 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.344497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.344559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.344580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.344614 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.344668 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.447922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.447970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.447987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.448009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.448026 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.551197 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.551256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.551270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.551290 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.551305 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.653538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.653575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.653584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.653598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.653607 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.755934 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.755970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.755980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.755996 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.756009 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.859144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.859191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.859203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.859222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.859236 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.961821 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.961882 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.961896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.961915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:30 crc kubenswrapper[4945]: I1008 15:14:30.961927 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:30Z","lastTransitionTime":"2025-10-08T15:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.063778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.063822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.063830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.063846 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.063857 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.166598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.166630 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.166637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.166652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.166661 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.268957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.269005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.269018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.269034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.269047 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.371658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.371718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.371733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.371755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.371772 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.474792 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.474854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.474871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.474896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.474915 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.480729 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/0.log" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.480814 4945 generic.go:334] "Generic (PLEG): container finished" podID="22e38fb8-8588-494f-a6cb-8b3d2c3a3142" containerID="e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140" exitCode=1 Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.480864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerDied","Data":"e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.481573 4945 scope.go:117] "RemoveContainer" containerID="e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.499180 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.525409 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:30Z\\\",\\\"message\\\":\\\"2025-10-08T15:13:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd\\\\n2025-10-08T15:13:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd to /host/opt/cni/bin/\\\\n2025-10-08T15:13:45Z [verbose] multus-daemon started\\\\n2025-10-08T15:13:45Z [verbose] Readiness Indicator file check\\\\n2025-10-08T15:14:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.546514 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.564872 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.577652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.577691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.577717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.577733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.577743 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.580166 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.601469 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.623001 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.638069 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.649989 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.671857 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.680054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.680092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.680107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.680144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.680161 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.686897 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.701526 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.714764 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.729888 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.743664 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.761444 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.774639 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.781782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.781839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.781854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.781872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.781883 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.789393 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:31Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.884055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.884111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.884134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.884156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.884169 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.987810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.988389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.988401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.988424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:31 crc kubenswrapper[4945]: I1008 15:14:31.988438 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:31Z","lastTransitionTime":"2025-10-08T15:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.023247 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.023293 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.023466 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.023452 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.023662 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.023680 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.024023 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.024529 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.024938 4945 scope.go:117] "RemoveContainer" containerID="4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.025289 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.041412 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.060251 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.076158 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.089835 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.091244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.091298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.091311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.091339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.091355 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.102501 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.118494 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.132323 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.146711 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:30Z\\\",\\\"message\\\":\\\"2025-10-08T15:13:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd\\\\n2025-10-08T15:13:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd to /host/opt/cni/bin/\\\\n2025-10-08T15:13:45Z [verbose] multus-daemon started\\\\n2025-10-08T15:13:45Z [verbose] Readiness Indicator file check\\\\n2025-10-08T15:14:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.168769 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.185903 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.193923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.193949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.193958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.193973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.193982 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.207169 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.223490 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.237240 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.244213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.244248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.244258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.244273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.244282 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.251960 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.255278 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.258209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.258244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.258253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.258266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.258276 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.262313 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.269463 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.272537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.272560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.272569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.272584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.272593 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.276443 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.285671 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.289193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.289238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.289253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.289274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.289291 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.289312 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.299558 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.301517 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.302730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.302759 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.302767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.302780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.302790 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.318546 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: E1008 15:14:32.319022 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.320858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.320936 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.320958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.320985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.321002 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.423588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.423639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.423649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.423663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.423672 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.485594 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/0.log" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.485673 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerStarted","Data":"892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.500428 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.515950 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.526445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.526510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.526527 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.526554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.526574 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.532936 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.553760 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.570410 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.582047 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.593728 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.606639 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.622504 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.629993 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.630033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.630044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.630061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.630071 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.636671 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.646898 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.661216 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.674352 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.687263 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:30Z\\\",\\\"message\\\":\\\"2025-10-08T15:13:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd\\\\n2025-10-08T15:13:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd to /host/opt/cni/bin/\\\\n2025-10-08T15:13:45Z [verbose] multus-daemon started\\\\n2025-10-08T15:13:45Z [verbose] Readiness Indicator file check\\\\n2025-10-08T15:14:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.712961 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.724967 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.732865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.732912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.732924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.732938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.732950 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.734829 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.743885 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:32Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.835871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.835919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.835934 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.835954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.835968 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.939013 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.939060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.939072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.939089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:32 crc kubenswrapper[4945]: I1008 15:14:32.939101 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:32Z","lastTransitionTime":"2025-10-08T15:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.041874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.041923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.041933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.041952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.041966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.144922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.144960 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.144969 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.144983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.144992 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.247634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.247706 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.247716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.247733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.247744 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.349922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.350080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.350103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.350177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.350197 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.452450 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.452501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.452518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.452541 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.452560 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.555251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.555331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.555357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.555551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.555594 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.658954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.658996 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.659007 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.659024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.659037 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.761657 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.761731 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.761749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.761774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.761793 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.864301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.864328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.864337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.864349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.864358 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.967657 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.968000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.968180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.968318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:33 crc kubenswrapper[4945]: I1008 15:14:33.968434 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:33Z","lastTransitionTime":"2025-10-08T15:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.024197 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.024266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.024336 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:34 crc kubenswrapper[4945]: E1008 15:14:34.024497 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.024571 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:34 crc kubenswrapper[4945]: E1008 15:14:34.024703 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:34 crc kubenswrapper[4945]: E1008 15:14:34.024726 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:34 crc kubenswrapper[4945]: E1008 15:14:34.024855 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.071326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.071371 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.071403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.071421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.071433 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.175491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.175583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.175599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.175647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.175662 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.285524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.285952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.286209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.286384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.286570 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.390196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.390271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.390293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.390321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.390339 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.492509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.492578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.492605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.492637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.492659 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.595960 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.596043 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.596061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.596086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.596104 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.698917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.698979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.698995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.699017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.699031 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.801645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.801742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.801761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.801784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.801801 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.905364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.905440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.905472 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.905492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:34 crc kubenswrapper[4945]: I1008 15:14:34.905506 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:34Z","lastTransitionTime":"2025-10-08T15:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.008586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.008874 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.008938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.008999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.009158 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.112647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.112692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.112708 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.112729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.112746 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.215854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.215933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.215956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.215988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.216013 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.318386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.318449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.318473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.318504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.318527 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.421509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.421549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.421557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.421573 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.421582 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.524885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.524967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.524990 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.525014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.525034 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.627484 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.627544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.627560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.627585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.627603 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.730314 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.730391 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.730403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.730420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.730432 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.833381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.833457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.833482 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.833515 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.833540 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.936413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.936461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.936476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.936497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:35 crc kubenswrapper[4945]: I1008 15:14:35.936510 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:35Z","lastTransitionTime":"2025-10-08T15:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.024033 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.024165 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:36 crc kubenswrapper[4945]: E1008 15:14:36.024259 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.024069 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:36 crc kubenswrapper[4945]: E1008 15:14:36.024452 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.024538 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:36 crc kubenswrapper[4945]: E1008 15:14:36.024579 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:36 crc kubenswrapper[4945]: E1008 15:14:36.024786 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.039347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.039376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.039386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.039399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.039408 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.140687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.140750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.140827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.140845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.140855 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.243317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.243358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.243379 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.243399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.243413 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.346320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.346376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.346394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.346422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.346442 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.450223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.450297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.450319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.450347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.450367 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.553571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.553693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.553728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.553762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.553784 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.665902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.665985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.666011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.666038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.666056 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.769224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.769301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.769325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.769356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.769378 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.872578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.872778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.872824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.872863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.872888 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.975519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.975593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.975616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.975643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:36 crc kubenswrapper[4945]: I1008 15:14:36.975659 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:36Z","lastTransitionTime":"2025-10-08T15:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.078257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.078340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.078363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.078393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.078416 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.181849 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.181924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.181963 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.181995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.182021 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.284877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.284957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.284985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.285001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.285011 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.388186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.388256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.388280 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.388307 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.388326 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.491192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.491256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.491273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.491300 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.491319 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.593941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.594000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.594017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.594040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.594058 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.696847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.696886 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.696894 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.696906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.696916 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.768538 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.768787 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:41.768753302 +0000 UTC m=+151.122668233 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.769403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.769612 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.769721 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:15:41.769696307 +0000 UTC m=+151.123611248 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.799926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.799963 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.799973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.799990 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.800000 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.870974 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.871473 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.871692 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.871199 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.871597 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.872163 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.872180 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.871769 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.872217 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.872226 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.872622 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:15:41.872085196 +0000 UTC m=+151.226000147 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.872817 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:15:41.872795624 +0000 UTC m=+151.226710565 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:37 crc kubenswrapper[4945]: E1008 15:14:37.873019 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:15:41.873000629 +0000 UTC m=+151.226915570 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.902712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.902967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.903175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.903350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:37 crc kubenswrapper[4945]: I1008 15:14:37.903487 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:37Z","lastTransitionTime":"2025-10-08T15:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.006402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.006460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.006477 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.006501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.006518 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.024053 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.024053 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.024171 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.024142 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:38 crc kubenswrapper[4945]: E1008 15:14:38.024239 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:38 crc kubenswrapper[4945]: E1008 15:14:38.024347 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:38 crc kubenswrapper[4945]: E1008 15:14:38.024450 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:38 crc kubenswrapper[4945]: E1008 15:14:38.024550 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.109398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.109461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.109471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.109490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.109500 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.211976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.212016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.212032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.212051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.212065 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.315211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.315278 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.315304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.315336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.315363 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.419217 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.419274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.419293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.419315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.419331 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.522191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.522265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.522283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.522312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.522333 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.625477 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.626268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.626304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.626331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.626353 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.729949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.730016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.730035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.730060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.730078 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.833377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.833436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.833448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.833466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.833479 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.936382 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.936455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.936467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.936486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:38 crc kubenswrapper[4945]: I1008 15:14:38.936498 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:38Z","lastTransitionTime":"2025-10-08T15:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.040103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.040181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.040193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.040212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.040230 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.143469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.143588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.143619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.143704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.143725 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.246907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.246997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.247025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.247061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.247103 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.350822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.350885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.350905 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.350929 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.350947 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.454562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.454629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.454646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.454674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.454699 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.558474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.558544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.558568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.558597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.558618 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.662342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.662429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.662452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.662487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.662513 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.764970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.765009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.765018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.765034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.765045 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.869577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.869667 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.869686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.869717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.869738 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.973153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.973231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.973250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.973291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:39 crc kubenswrapper[4945]: I1008 15:14:39.973309 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:39Z","lastTransitionTime":"2025-10-08T15:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.023798 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:40 crc kubenswrapper[4945]: E1008 15:14:40.023957 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.024265 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.024332 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:40 crc kubenswrapper[4945]: E1008 15:14:40.024426 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:40 crc kubenswrapper[4945]: E1008 15:14:40.024512 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.024576 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:40 crc kubenswrapper[4945]: E1008 15:14:40.024778 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.076737 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.077185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.077388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.077641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.077836 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.181939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.181984 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.181995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.182021 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.182043 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.285403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.285473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.285495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.285523 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.285545 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.387775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.387839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.387861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.387885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.387902 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.491724 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.491790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.491806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.491833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.491848 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.594685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.594739 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.594759 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.594783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.594802 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.697662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.697727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.697749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.697776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.697797 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.800308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.800356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.800364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.800378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.800387 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.902233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.902283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.902294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.902311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:40 crc kubenswrapper[4945]: I1008 15:14:40.902322 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:40Z","lastTransitionTime":"2025-10-08T15:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.004579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.004628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.004650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.004670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.004683 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.107647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.107693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.107707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.107729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.107744 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.210016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.210396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.210411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.210426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.210436 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.313629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.313941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.314080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.314209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.314322 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.417078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.417459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.417658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.417853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.417998 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.520705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.520752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.520768 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.520787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.520801 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.624154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.624210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.624227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.624251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.624268 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.726161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.726203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.726220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.726241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.726256 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.828099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.828162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.828173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.828189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.828200 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.931006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.931051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.931062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.931078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:41 crc kubenswrapper[4945]: I1008 15:14:41.931087 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:41Z","lastTransitionTime":"2025-10-08T15:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.023734 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.023910 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.023977 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.024172 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.024553 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.024928 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.025200 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.025349 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.034405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.034467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.034489 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.034514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.034532 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.047572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.064997 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.081609 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.095691 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.109947 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.136848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.136889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.136898 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.136916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.136940 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.138898 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.151431 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.161863 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.172378 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.184579 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:30Z\\\",\\\"message\\\":\\\"2025-10-08T15:13:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd\\\\n2025-10-08T15:13:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd to /host/opt/cni/bin/\\\\n2025-10-08T15:13:45Z [verbose] multus-daemon started\\\\n2025-10-08T15:13:45Z [verbose] Readiness Indicator file check\\\\n2025-10-08T15:14:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.203556 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.216734 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.236742 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.238610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.238675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.238684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.238719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.238730 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.250731 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.262000 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.272033 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.283212 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.293554 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.341372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.341414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.341425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.341440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.341450 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.411533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.411754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.411997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.412180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.412352 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.422632 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.425940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.425974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.425983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.425997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.426009 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.437405 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.440947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.440988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.441003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.441020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.441034 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.453216 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.456299 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.456330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.456341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.456359 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.456373 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.472484 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.476207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.476251 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.476261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.476289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.476300 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.487672 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"907ff440-3709-4e53-8f2c-af6fb57f351a\\\",\\\"systemUUID\\\":\\\"c21442f4-cc84-4781-8fa8-fec367812c32\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:42Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:42 crc kubenswrapper[4945]: E1008 15:14:42.487775 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.489052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.489075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.489090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.489125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.489137 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.590680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.590727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.590741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.590761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.590775 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.693699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.693776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.693824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.693848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.693862 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.797644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.797718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.797742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.797778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.797802 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.906650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.906728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.906748 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.906776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:42 crc kubenswrapper[4945]: I1008 15:14:42.906795 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:42Z","lastTransitionTime":"2025-10-08T15:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.010415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.010483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.010505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.010532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.010553 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.112993 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.113062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.113075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.113093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.113122 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.215249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.215508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.215517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.215530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.215540 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.344003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.344045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.344056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.344072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.344085 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.446540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.446708 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.446726 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.446785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.446803 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.549656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.549705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.549717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.549733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.549744 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.652658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.652718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.652735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.652757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.652775 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.755884 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.755944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.755962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.755988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.756010 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.858504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.858536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.858549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.858563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.858574 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.960977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.961013 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.961022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.961035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:43 crc kubenswrapper[4945]: I1008 15:14:43.961044 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:43Z","lastTransitionTime":"2025-10-08T15:14:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.023744 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.023782 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:44 crc kubenswrapper[4945]: E1008 15:14:44.024051 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.024340 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.024338 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:44 crc kubenswrapper[4945]: E1008 15:14:44.024673 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:44 crc kubenswrapper[4945]: E1008 15:14:44.024948 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:44 crc kubenswrapper[4945]: E1008 15:14:44.025060 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.064135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.064220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.064240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.064266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.064281 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.166511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.166582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.166600 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.166624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.166640 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.268806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.268872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.268896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.268926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.268949 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.371752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.371810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.371831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.371854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.371871 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.474987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.475065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.475089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.475155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.475181 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.577800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.577836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.577845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.577859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.577869 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.680693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.680778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.680804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.680837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.680860 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.784174 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.784232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.784249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.784273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.784294 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.887288 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.887351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.887364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.887381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.887395 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.991282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.991387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.991405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.991435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:44 crc kubenswrapper[4945]: I1008 15:14:44.991455 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:44Z","lastTransitionTime":"2025-10-08T15:14:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.024925 4945 scope.go:117] "RemoveContainer" containerID="4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.035419 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.094420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.094492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.094516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.094550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.094575 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.197911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.197946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.197959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.197976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.197991 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.300050 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.300095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.300104 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.300147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.300161 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.402871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.402932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.402949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.402974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.402992 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.504803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.504837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.504845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.504859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.504868 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.531932 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/2.log" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.534733 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.547984 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.559909 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.568278 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0aadf286-76e0-49d6-8baa-f987a11e73ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8056d279f7555fd7f6ab9350a8550760e8fab4de9d918f9663c2285595d752fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7485ffbe62529c9a2f879906870ca3bc5b96d7a2fc704a8c722eb99ddd81bd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7485ffbe62529c9a2f879906870ca3bc5b96d7a2fc704a8c722eb99ddd81bd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.578146 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.590658 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.600497 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.607099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.607155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.607163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.607178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.607187 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.611069 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.621599 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.638164 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:30Z\\\",\\\"message\\\":\\\"2025-10-08T15:13:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd\\\\n2025-10-08T15:13:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd to /host/opt/cni/bin/\\\\n2025-10-08T15:13:45Z [verbose] multus-daemon started\\\\n2025-10-08T15:13:45Z [verbose] Readiness Indicator file check\\\\n2025-10-08T15:14:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.657271 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.673745 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.701568 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.709025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.709055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.709063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.709076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.709085 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.721277 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.735963 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.750226 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.759685 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.770808 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.783301 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.792481 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:45Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.811537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.811568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.811577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.811593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.811606 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.914330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.914374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.914390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.914413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:45 crc kubenswrapper[4945]: I1008 15:14:45.914429 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:45Z","lastTransitionTime":"2025-10-08T15:14:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.017027 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.017165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.017192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.017229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.017252 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.024080 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.024229 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.024101 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:46 crc kubenswrapper[4945]: E1008 15:14:46.024301 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.024326 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:46 crc kubenswrapper[4945]: E1008 15:14:46.024478 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:46 crc kubenswrapper[4945]: E1008 15:14:46.024590 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:46 crc kubenswrapper[4945]: E1008 15:14:46.024725 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.120707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.121085 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.121142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.121176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.121200 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.223630 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.223687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.223696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.223735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.223745 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.326764 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.326837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.326850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.326867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.326880 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.429770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.429827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.429843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.429864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.429879 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.532415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.532467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.532485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.532508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.532526 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.541246 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/3.log" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.542333 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/2.log" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.547091 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" exitCode=1 Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.547143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.547224 4945 scope.go:117] "RemoveContainer" containerID="4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.549870 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:14:46 crc kubenswrapper[4945]: E1008 15:14:46.550156 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.564900 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dpk49\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zj7s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.596934 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dab13642-62c1-4c20-8273-80e49bffdd9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d621822258b21988748b227209e2479d01d98c4b5779031d3a95e88adc1c13d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab444ce93c8bd0e467857fdf768cfe26c5a17228edad920d2d400026e965d31b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a554fd3e1bc3119155a895d78307bea70a45237a1f0b57a856a3f908318aed9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416310934e762178ed442ac5ce51952168c002ac90fa88fc9fa6fad3c4c3d528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49c1f00fda36cc5de2cd2d34a73516efc774d6657193b75b9d5fbc73310b65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb5ea726739e85785b63bc5989b8f7fc43bc72b303a5f71a824de3fff45bf17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99aa3698381eabe564f8c46de11dd3584bc892d9445f8a9dc2f86b9651c0e0d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b909b3fdadd8e055d79708a03d781033e94e34d3201af4f78905db878ea7dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.618336 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a039691b-bc35-4f58-9694-3e60c7252bdc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a59afb7a96c35e679ec49a2ab338bdcaf0185243018813f4e7b242a5aa7338e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bb11548568ba358b3f2ed3795d5226a8338251190bc6bed03849641123be8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f94f594f53dd3769ca12bad550cb40e85e6cab8ec0dda372e9b2f0195cbb93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://062c198a4e2d6e2ad7cc28cf9cc20545870bc87cbd3eb957f797bd6a91ae4e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f1f10be250f2116a82a0622124dbc6fdf4f05dcc18e02fcead2601a72a09dbf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T15:13:34Z\\\",\\\"message\\\":\\\"le observer\\\\nW1008 15:13:33.722879 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1008 15:13:33.730801 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 15:13:33.731570 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2330779153/tls.crt::/tmp/serving-cert-2330779153/tls.key\\\\\\\"\\\\nI1008 15:13:34.175405 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 15:13:34.182432 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 15:13:34.182471 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 15:13:34.182512 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 15:13:34.182523 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 15:13:34.191498 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 15:13:34.191525 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191534 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 15:13:34.191541 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 15:13:34.191545 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 15:13:34.191548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 15:13:34.191552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 15:13:34.191597 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 15:13:34.194398 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d581232e5c4004cce3e7fa546b4d6a8e7e77479d399e3ccb262100a31608599\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79ba55fa1ecc77de99f2adad3044c74840523c5de946090a54a730494048b6dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.634407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.634443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.634454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.634470 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.634481 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.635230 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.656420 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e93c84e1-87ae-4485-8ea1-8cd4ac1554a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61cf593dfdee1281abcb40570dbab23b2364f942d87ce61f8e6b20a891e408c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88270e850df9d60664906eb4273c59e501ac44f8926b92b78723679789739259\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c32243dfca4444c0c0dd46a6a7fd7b2e97d6a86fe173efc26e175c2355e0ba6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d66b60c9126d507a5e4cb54c067db683237453b6e9727d936cc5b6209460bf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d461e0e3fa645019a7bf8940f411a739250368e498d846b89b1edde5577b6ccc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6143239ea17519700699f459c2ff837a0e93a865f40d593ce2aa38a68dd14cfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bde04a512c4fc748f1970ad410de654d1b36bc5af8b7c931bb9a444035454f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mzlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2gdgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.668813 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5317b90a1f93698c45df34b4a31a9fdd37fded9aac18ae8d506d54dd7bd19ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f67093496da49fe08f4230b2deaf7500f672b58bd93343fedb594706e0258e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.688981 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67a2fd05d6b325e8641c1e80a293c57704211311480ec3432dfd643547a45f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.704631 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e39fc950-7f86-4d6b-b4b0-25daafb66250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90881a80af8b44eb374cb8992fe102153056358efc8c201405943abf8917769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njl6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jp64g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.715273 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.725589 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qb4pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4138a218-2d22-4042-ae74-5d4712982025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475d26d8c7cf17063f8dc00c5304ebf20e967d15e1826d352dffee018488036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkfnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qb4pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.736772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.736967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.737064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.737215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.737326 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.740622 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c0d0ee3-e2d0-462f-b3d3-14bec1defb2e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5a433cf8222b04d022dbcb94103e6fe0ea601bc1f67345daf899e1dbd106848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fce7a8911a6c9794976155f58b6817b758d84b4c95ea935ff314124bd4c36c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8283d7571155103125016a508fd373a03e532126df7a2f55a5e7fffd54e60c6d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5944fb585b83e9f565dc687937b41201f300156eead9622f6be05c740e9a05a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.753066 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95827f0-0c10-42e0-9982-9d2658d6e8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f9df1f52f6018c8d68882265c016aaa69a6c4ef10e61d7cfe4d4269adf692b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49d96bb660f4bce9954c81ef9b2894dd92aa4f8b4339e62dd8f0bb3509cdabb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db75e904e59c84250d7741fbdcb63e867777c1a439b298699cb4e5b6d7568198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee52acee9871824815dd3da524e34b1c909336b4505f903c4fd4c6372e7d12d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.763004 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0aadf286-76e0-49d6-8baa-f987a11e73ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8056d279f7555fd7f6ab9350a8550760e8fab4de9d918f9663c2285595d752fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7485ffbe62529c9a2f879906870ca3bc5b96d7a2fc704a8c722eb99ddd81bd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7485ffbe62529c9a2f879906870ca3bc5b96d7a2fc704a8c722eb99ddd81bd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.776370 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.790589 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55656f9e-8162-446e-8b48-5947f35613f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53e3b740c4769c2a1ba0f3bfd670c648f83d8412fa39cc5277866ce64882cc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda9a41b81fac9f321b039f420d21e1cd4c21abc0e0b86e028140dddeb9519d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv5l5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p66bn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.805100 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e30e884415f1c51b2290724e209efcd37390b4e271f1628f078388c80f4708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.816232 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z7jwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7a1789a-2045-4ec7-884e-eb70c9031197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41b4f06dcf4adc2090ad21ef6ec93c1ca9b5b5f5fa826b498830b0098aadeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xq6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:40Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z7jwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.829580 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8kc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22e38fb8-8588-494f-a6cb-8b3d2c3a3142\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:30Z\\\",\\\"message\\\":\\\"2025-10-08T15:13:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd\\\\n2025-10-08T15:13:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4525502b-1bd9-4535-ac78-68f37d3cb1dd to /host/opt/cni/bin/\\\\n2025-10-08T15:13:45Z [verbose] multus-daemon started\\\\n2025-10-08T15:13:45Z [verbose] Readiness Indicator file check\\\\n2025-10-08T15:14:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rksk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8kc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.840077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.840168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.840188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.840205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.840215 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.850695 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28239584-598a-49d2-a9b0-189e4f4ad733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T15:13:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b39ce9d98b540a1c358a4551ad256b4416f76d3119a01c6ef7526797c377abf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:16Z\\\",\\\"message\\\":\\\"5:14:15.797030 6751 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1008 15:14:15.797063 6751 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1008 15:14:15.797133 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 15:14:15.797134 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T15:14:45Z\\\",\\\"message\\\":\\\" github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 15:14:45.827401 7150 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 15:14:45.829712 7150 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 15:14:45.830068 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 15:14:45.830133 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 15:14:45.830155 7150 factory.go:656] Stopping watch factory\\\\nI1008 15:14:45.830180 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 15:14:45.830187 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 15:14:45.831292 7150 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1008 15:14:45.831307 7150 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1008 15:14:45.831339 7150 ovnkube.go:599] Stopped ovnkube\\\\nI1008 15:14:45.831355 7150 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 15:14:45.831453 7150 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T15:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T15:13:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T15:13:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T15:13:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fzlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T15:13:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9g2vg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T15:14:46Z is after 2025-08-24T17:21:41Z" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.943879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.943937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.943954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.943979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:46 crc kubenswrapper[4945]: I1008 15:14:46.943997 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:46Z","lastTransitionTime":"2025-10-08T15:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.047440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.047491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.047506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.047526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.047539 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.150224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.150272 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.150289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.150311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.150326 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.253532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.253594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.253618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.253647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.253670 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.357439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.357481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.357492 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.357510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.357522 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.460461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.460518 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.460535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.460560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.460576 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.552232 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/3.log" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.561741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.561775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.561786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.561800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.561810 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.664403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.664455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.664486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.664506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.664519 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.766901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.766946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.766957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.766977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.766990 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.869677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.869707 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.869715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.869729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.869737 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.972539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.972618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.972644 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.972677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:47 crc kubenswrapper[4945]: I1008 15:14:47.972700 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:47Z","lastTransitionTime":"2025-10-08T15:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.023453 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.023560 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.023573 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.023495 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:48 crc kubenswrapper[4945]: E1008 15:14:48.023704 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:48 crc kubenswrapper[4945]: E1008 15:14:48.023816 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:48 crc kubenswrapper[4945]: E1008 15:14:48.023916 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:48 crc kubenswrapper[4945]: E1008 15:14:48.024005 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.075825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.075918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.075942 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.075971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.075995 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.178105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.178152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.178160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.178173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.178182 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.280706 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.280757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.280769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.280787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.280802 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.383151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.383193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.383202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.383214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.383223 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.485301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.485368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.485429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.485453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.485471 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.587988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.588034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.588042 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.588057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.588067 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.691026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.691226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.691246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.691264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.691273 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.793231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.793283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.793293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.793309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.793318 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.896126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.896166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.896176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.896191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.896202 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.998912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.998956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.998966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.998992 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:48 crc kubenswrapper[4945]: I1008 15:14:48.999005 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:48Z","lastTransitionTime":"2025-10-08T15:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.101192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.101235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.101246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.101261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.101274 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.203466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.203541 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.203566 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.203595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.203615 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.307353 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.307439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.307449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.307472 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.307486 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.411317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.411508 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.411528 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.411555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.411573 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.514388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.514769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.514783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.514808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.514820 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.616987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.617033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.617053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.617075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.617089 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.719090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.719147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.719157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.719173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.719185 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.821631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.821687 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.821703 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.821725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.821741 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.924414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.924496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.924535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.924572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:49 crc kubenswrapper[4945]: I1008 15:14:49.924595 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:49Z","lastTransitionTime":"2025-10-08T15:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.023882 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.023957 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:50 crc kubenswrapper[4945]: E1008 15:14:50.024052 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.024064 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.024100 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:50 crc kubenswrapper[4945]: E1008 15:14:50.024188 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:50 crc kubenswrapper[4945]: E1008 15:14:50.024264 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:50 crc kubenswrapper[4945]: E1008 15:14:50.024573 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.026537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.026565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.026582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.026596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.026607 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.128225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.128262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.128270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.128284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.128293 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.231425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.231470 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.231484 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.231505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.231516 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.334725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.334786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.334803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.334827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.334844 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.437918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.438018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.438041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.438075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.438095 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.541098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.541156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.541165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.541181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.541192 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.643655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.643695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.643703 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.643723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.643732 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.747428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.747466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.747479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.747493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.747517 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.850556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.850796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.850890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.851019 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.851162 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.954974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.955270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.955366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.955451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:50 crc kubenswrapper[4945]: I1008 15:14:50.955568 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:50Z","lastTransitionTime":"2025-10-08T15:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.059207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.059276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.059313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.059341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.059360 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.163300 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.163353 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.163370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.163395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.163413 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.266359 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.266421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.266439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.266463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.266480 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.369297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.369371 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.369389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.369418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.369439 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.471864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.471945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.471970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.472002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.472026 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.574878 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.574918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.574928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.574945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.574957 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.677189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.677262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.677276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.677295 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.677306 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.780533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.780572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.780580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.780595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.780604 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.882872 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.882919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.882930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.882946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.882959 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.985743 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.985790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.985800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.985817 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:51 crc kubenswrapper[4945]: I1008 15:14:51.985827 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:51Z","lastTransitionTime":"2025-10-08T15:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.023242 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.023339 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:52 crc kubenswrapper[4945]: E1008 15:14:52.023362 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:52 crc kubenswrapper[4945]: E1008 15:14:52.023551 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.023635 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.023569 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:52 crc kubenswrapper[4945]: E1008 15:14:52.023789 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:52 crc kubenswrapper[4945]: E1008 15:14:52.023841 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.087968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.088018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.088032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.088051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.088063 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.091831 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podStartSLOduration=72.091815205 podStartE2EDuration="1m12.091815205s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.091434186 +0000 UTC m=+101.445349087" watchObservedRunningTime="2025-10-08 15:14:52.091815205 +0000 UTC m=+101.445730106" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.116094 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.116074376 podStartE2EDuration="1m19.116074376s" podCreationTimestamp="2025-10-08 15:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.104021547 +0000 UTC m=+101.457936448" watchObservedRunningTime="2025-10-08 15:14:52.116074376 +0000 UTC m=+101.469989277" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.129609 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=49.129588271 podStartE2EDuration="49.129588271s" podCreationTimestamp="2025-10-08 15:14:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.116727502 +0000 UTC m=+101.470642393" watchObservedRunningTime="2025-10-08 15:14:52.129588271 +0000 UTC m=+101.483503162" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.146208 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=7.146183136 podStartE2EDuration="7.146183136s" podCreationTimestamp="2025-10-08 15:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.130011662 +0000 UTC m=+101.483926593" watchObservedRunningTime="2025-10-08 15:14:52.146183136 +0000 UTC m=+101.500098037" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.178363 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-qb4pj" podStartSLOduration=72.178338218 podStartE2EDuration="1m12.178338218s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.177800625 +0000 UTC m=+101.531715536" watchObservedRunningTime="2025-10-08 15:14:52.178338218 +0000 UTC m=+101.532253129" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.190650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.190693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.190709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.190731 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.190748 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.202206 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-z7jwc" podStartSLOduration=72.202190108 podStartE2EDuration="1m12.202190108s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.201979113 +0000 UTC m=+101.555894024" watchObservedRunningTime="2025-10-08 15:14:52.202190108 +0000 UTC m=+101.556105029" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.236751 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-j8kc9" podStartSLOduration=72.236733502 podStartE2EDuration="1m12.236733502s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.216258478 +0000 UTC m=+101.570173379" watchObservedRunningTime="2025-10-08 15:14:52.236733502 +0000 UTC m=+101.590648403" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.257472 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p66bn" podStartSLOduration=72.257458942 podStartE2EDuration="1m12.257458942s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.256743244 +0000 UTC m=+101.610658155" watchObservedRunningTime="2025-10-08 15:14:52.257458942 +0000 UTC m=+101.611373843" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.293417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.293448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.293457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.293469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.293478 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.295463 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=77.295449334 podStartE2EDuration="1m17.295449334s" podCreationTimestamp="2025-10-08 15:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.280012689 +0000 UTC m=+101.633927610" watchObservedRunningTime="2025-10-08 15:14:52.295449334 +0000 UTC m=+101.649364235" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.295676 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.29567101 podStartE2EDuration="1m17.29567101s" podCreationTimestamp="2025-10-08 15:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.293275268 +0000 UTC m=+101.647190169" watchObservedRunningTime="2025-10-08 15:14:52.29567101 +0000 UTC m=+101.649585911" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.320979 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2gdgm" podStartSLOduration=72.320949166 podStartE2EDuration="1m12.320949166s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:52.320648139 +0000 UTC m=+101.674563060" watchObservedRunningTime="2025-10-08 15:14:52.320949166 +0000 UTC m=+101.674864067" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.396164 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.396210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.396224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.396241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.396288 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.498907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.498968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.499001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.499033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.499056 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.603025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.603179 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.603200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.603235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.603258 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.706183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.706252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.706280 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.706309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.706332 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.746618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.746662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.746673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.746689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.746698 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T15:14:52Z","lastTransitionTime":"2025-10-08T15:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.809341 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22"] Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.809753 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.811316 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.811935 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.812774 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.812862 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.955852 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/632edb79-89c7-43f2-a186-a5565f38a402-service-ca\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.955894 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/632edb79-89c7-43f2-a186-a5565f38a402-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.955928 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/632edb79-89c7-43f2-a186-a5565f38a402-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.955951 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/632edb79-89c7-43f2-a186-a5565f38a402-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:52 crc kubenswrapper[4945]: I1008 15:14:52.956026 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/632edb79-89c7-43f2-a186-a5565f38a402-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.057288 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/632edb79-89c7-43f2-a186-a5565f38a402-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.057349 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/632edb79-89c7-43f2-a186-a5565f38a402-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.057370 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/632edb79-89c7-43f2-a186-a5565f38a402-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.057426 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/632edb79-89c7-43f2-a186-a5565f38a402-service-ca\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.057460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/632edb79-89c7-43f2-a186-a5565f38a402-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.057471 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/632edb79-89c7-43f2-a186-a5565f38a402-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.057501 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/632edb79-89c7-43f2-a186-a5565f38a402-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.058722 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/632edb79-89c7-43f2-a186-a5565f38a402-service-ca\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.063910 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/632edb79-89c7-43f2-a186-a5565f38a402-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.084217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/632edb79-89c7-43f2-a186-a5565f38a402-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-97g22\" (UID: \"632edb79-89c7-43f2-a186-a5565f38a402\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.122654 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.574228 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" event={"ID":"632edb79-89c7-43f2-a186-a5565f38a402","Type":"ContainerStarted","Data":"b65b423046006e81d7cf443df9a86dd82b7db1f9f3591fe84136edbb11df82a4"} Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.574286 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" event={"ID":"632edb79-89c7-43f2-a186-a5565f38a402","Type":"ContainerStarted","Data":"8c19bfc65b21ed7c785e2c07707307969b1ffc608ef288143ed9b0111d1a58c1"} Oct 08 15:14:53 crc kubenswrapper[4945]: I1008 15:14:53.589787 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-97g22" podStartSLOduration=73.589773089 podStartE2EDuration="1m13.589773089s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:14:53.589563014 +0000 UTC m=+102.943477945" watchObservedRunningTime="2025-10-08 15:14:53.589773089 +0000 UTC m=+102.943687990" Oct 08 15:14:54 crc kubenswrapper[4945]: I1008 15:14:54.023809 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:54 crc kubenswrapper[4945]: I1008 15:14:54.023944 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:54 crc kubenswrapper[4945]: E1008 15:14:54.024180 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:54 crc kubenswrapper[4945]: I1008 15:14:54.024203 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:54 crc kubenswrapper[4945]: E1008 15:14:54.023960 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:54 crc kubenswrapper[4945]: E1008 15:14:54.024339 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:54 crc kubenswrapper[4945]: I1008 15:14:54.024519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:54 crc kubenswrapper[4945]: E1008 15:14:54.024591 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:55 crc kubenswrapper[4945]: I1008 15:14:55.892873 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:14:55 crc kubenswrapper[4945]: I1008 15:14:55.893930 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:14:55 crc kubenswrapper[4945]: E1008 15:14:55.894148 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:14:56 crc kubenswrapper[4945]: I1008 15:14:56.023255 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:56 crc kubenswrapper[4945]: I1008 15:14:56.023325 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:56 crc kubenswrapper[4945]: I1008 15:14:56.023373 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:56 crc kubenswrapper[4945]: I1008 15:14:56.023281 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:56 crc kubenswrapper[4945]: E1008 15:14:56.023494 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:56 crc kubenswrapper[4945]: E1008 15:14:56.023557 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:56 crc kubenswrapper[4945]: E1008 15:14:56.023663 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:56 crc kubenswrapper[4945]: E1008 15:14:56.023768 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:58 crc kubenswrapper[4945]: I1008 15:14:58.023531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:58 crc kubenswrapper[4945]: I1008 15:14:58.023533 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:14:58 crc kubenswrapper[4945]: E1008 15:14:58.023718 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:14:58 crc kubenswrapper[4945]: I1008 15:14:58.023743 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:14:58 crc kubenswrapper[4945]: E1008 15:14:58.023973 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:14:58 crc kubenswrapper[4945]: I1008 15:14:58.024237 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:14:58 crc kubenswrapper[4945]: E1008 15:14:58.024413 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:14:58 crc kubenswrapper[4945]: E1008 15:14:58.024580 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:14:58 crc kubenswrapper[4945]: I1008 15:14:58.544730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:14:58 crc kubenswrapper[4945]: E1008 15:14:58.544900 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:14:58 crc kubenswrapper[4945]: E1008 15:14:58.545000 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs podName:6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0 nodeName:}" failed. No retries permitted until 2025-10-08 15:16:02.544978492 +0000 UTC m=+171.898893473 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs") pod "network-metrics-daemon-zj7s7" (UID: "6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 15:15:00 crc kubenswrapper[4945]: I1008 15:15:00.023546 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:00 crc kubenswrapper[4945]: I1008 15:15:00.023640 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:00 crc kubenswrapper[4945]: E1008 15:15:00.023716 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:00 crc kubenswrapper[4945]: I1008 15:15:00.023785 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:00 crc kubenswrapper[4945]: E1008 15:15:00.023996 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:00 crc kubenswrapper[4945]: I1008 15:15:00.024283 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:00 crc kubenswrapper[4945]: E1008 15:15:00.024416 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:00 crc kubenswrapper[4945]: E1008 15:15:00.024610 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:02 crc kubenswrapper[4945]: I1008 15:15:02.023398 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:02 crc kubenswrapper[4945]: E1008 15:15:02.024595 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:02 crc kubenswrapper[4945]: I1008 15:15:02.024690 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:02 crc kubenswrapper[4945]: I1008 15:15:02.024778 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:02 crc kubenswrapper[4945]: I1008 15:15:02.024715 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:02 crc kubenswrapper[4945]: E1008 15:15:02.024920 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:02 crc kubenswrapper[4945]: E1008 15:15:02.025010 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:02 crc kubenswrapper[4945]: E1008 15:15:02.025085 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:04 crc kubenswrapper[4945]: I1008 15:15:04.024411 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:04 crc kubenswrapper[4945]: E1008 15:15:04.025072 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:04 crc kubenswrapper[4945]: I1008 15:15:04.024730 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:04 crc kubenswrapper[4945]: E1008 15:15:04.025231 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:04 crc kubenswrapper[4945]: I1008 15:15:04.024759 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:04 crc kubenswrapper[4945]: E1008 15:15:04.025331 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:04 crc kubenswrapper[4945]: I1008 15:15:04.024640 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:04 crc kubenswrapper[4945]: E1008 15:15:04.025414 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:06 crc kubenswrapper[4945]: I1008 15:15:06.023846 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:06 crc kubenswrapper[4945]: I1008 15:15:06.023842 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:06 crc kubenswrapper[4945]: I1008 15:15:06.023989 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:06 crc kubenswrapper[4945]: I1008 15:15:06.024056 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:06 crc kubenswrapper[4945]: E1008 15:15:06.024313 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:06 crc kubenswrapper[4945]: E1008 15:15:06.024463 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:06 crc kubenswrapper[4945]: E1008 15:15:06.024569 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:06 crc kubenswrapper[4945]: E1008 15:15:06.024664 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:08 crc kubenswrapper[4945]: I1008 15:15:08.023945 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:08 crc kubenswrapper[4945]: I1008 15:15:08.024081 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:08 crc kubenswrapper[4945]: E1008 15:15:08.024172 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:08 crc kubenswrapper[4945]: I1008 15:15:08.024198 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:08 crc kubenswrapper[4945]: I1008 15:15:08.024198 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:08 crc kubenswrapper[4945]: E1008 15:15:08.024342 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:08 crc kubenswrapper[4945]: E1008 15:15:08.024496 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:08 crc kubenswrapper[4945]: E1008 15:15:08.024633 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:10 crc kubenswrapper[4945]: I1008 15:15:10.023666 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:10 crc kubenswrapper[4945]: I1008 15:15:10.023764 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:10 crc kubenswrapper[4945]: I1008 15:15:10.023877 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:10 crc kubenswrapper[4945]: E1008 15:15:10.023899 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:10 crc kubenswrapper[4945]: I1008 15:15:10.023927 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:10 crc kubenswrapper[4945]: E1008 15:15:10.024023 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:10 crc kubenswrapper[4945]: E1008 15:15:10.024162 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:10 crc kubenswrapper[4945]: E1008 15:15:10.024281 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:11 crc kubenswrapper[4945]: I1008 15:15:11.025500 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:15:11 crc kubenswrapper[4945]: E1008 15:15:11.025799 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:15:12 crc kubenswrapper[4945]: I1008 15:15:12.024175 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:12 crc kubenswrapper[4945]: I1008 15:15:12.024266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:12 crc kubenswrapper[4945]: I1008 15:15:12.025411 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:12 crc kubenswrapper[4945]: I1008 15:15:12.025438 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:12 crc kubenswrapper[4945]: E1008 15:15:12.025379 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:12 crc kubenswrapper[4945]: E1008 15:15:12.025552 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:12 crc kubenswrapper[4945]: E1008 15:15:12.025754 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:12 crc kubenswrapper[4945]: E1008 15:15:12.025962 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:12 crc kubenswrapper[4945]: E1008 15:15:12.057464 4945 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 08 15:15:12 crc kubenswrapper[4945]: E1008 15:15:12.247995 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 15:15:14 crc kubenswrapper[4945]: I1008 15:15:14.024245 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:14 crc kubenswrapper[4945]: I1008 15:15:14.024297 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:14 crc kubenswrapper[4945]: E1008 15:15:14.024458 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:14 crc kubenswrapper[4945]: I1008 15:15:14.024523 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:14 crc kubenswrapper[4945]: I1008 15:15:14.024613 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:14 crc kubenswrapper[4945]: E1008 15:15:14.024815 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:14 crc kubenswrapper[4945]: E1008 15:15:14.024956 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:14 crc kubenswrapper[4945]: E1008 15:15:14.025022 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:16 crc kubenswrapper[4945]: I1008 15:15:16.023787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:16 crc kubenswrapper[4945]: I1008 15:15:16.023787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:16 crc kubenswrapper[4945]: I1008 15:15:16.024983 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:16 crc kubenswrapper[4945]: E1008 15:15:16.025201 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:16 crc kubenswrapper[4945]: I1008 15:15:16.025247 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:16 crc kubenswrapper[4945]: E1008 15:15:16.025327 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:16 crc kubenswrapper[4945]: E1008 15:15:16.025391 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:16 crc kubenswrapper[4945]: E1008 15:15:16.025479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:17 crc kubenswrapper[4945]: E1008 15:15:17.249816 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 15:15:17 crc kubenswrapper[4945]: I1008 15:15:17.662404 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/1.log" Oct 08 15:15:17 crc kubenswrapper[4945]: I1008 15:15:17.663671 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/0.log" Oct 08 15:15:17 crc kubenswrapper[4945]: I1008 15:15:17.663762 4945 generic.go:334] "Generic (PLEG): container finished" podID="22e38fb8-8588-494f-a6cb-8b3d2c3a3142" containerID="892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203" exitCode=1 Oct 08 15:15:17 crc kubenswrapper[4945]: I1008 15:15:17.663803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerDied","Data":"892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203"} Oct 08 15:15:17 crc kubenswrapper[4945]: I1008 15:15:17.663848 4945 scope.go:117] "RemoveContainer" containerID="e2a5803df05cb12e34d1ae58a9702eaaf1fd8c0a75f133bb5ba11436bffbd140" Oct 08 15:15:17 crc kubenswrapper[4945]: I1008 15:15:17.665568 4945 scope.go:117] "RemoveContainer" containerID="892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203" Oct 08 15:15:17 crc kubenswrapper[4945]: E1008 15:15:17.667051 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-j8kc9_openshift-multus(22e38fb8-8588-494f-a6cb-8b3d2c3a3142)\"" pod="openshift-multus/multus-j8kc9" podUID="22e38fb8-8588-494f-a6cb-8b3d2c3a3142" Oct 08 15:15:18 crc kubenswrapper[4945]: I1008 15:15:18.023613 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:18 crc kubenswrapper[4945]: I1008 15:15:18.023657 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:18 crc kubenswrapper[4945]: I1008 15:15:18.023718 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:18 crc kubenswrapper[4945]: E1008 15:15:18.024218 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:18 crc kubenswrapper[4945]: I1008 15:15:18.024366 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:18 crc kubenswrapper[4945]: E1008 15:15:18.024371 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:18 crc kubenswrapper[4945]: E1008 15:15:18.024062 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:18 crc kubenswrapper[4945]: E1008 15:15:18.024781 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:18 crc kubenswrapper[4945]: I1008 15:15:18.668515 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/1.log" Oct 08 15:15:20 crc kubenswrapper[4945]: I1008 15:15:20.023977 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:20 crc kubenswrapper[4945]: I1008 15:15:20.024041 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:20 crc kubenswrapper[4945]: I1008 15:15:20.024079 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:20 crc kubenswrapper[4945]: I1008 15:15:20.023976 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:20 crc kubenswrapper[4945]: E1008 15:15:20.024190 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:20 crc kubenswrapper[4945]: E1008 15:15:20.024333 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:20 crc kubenswrapper[4945]: E1008 15:15:20.024446 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:20 crc kubenswrapper[4945]: E1008 15:15:20.024615 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:22 crc kubenswrapper[4945]: I1008 15:15:22.023238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:22 crc kubenswrapper[4945]: I1008 15:15:22.023262 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:22 crc kubenswrapper[4945]: I1008 15:15:22.023448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:22 crc kubenswrapper[4945]: I1008 15:15:22.023542 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:22 crc kubenswrapper[4945]: E1008 15:15:22.024362 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:22 crc kubenswrapper[4945]: E1008 15:15:22.024436 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:22 crc kubenswrapper[4945]: E1008 15:15:22.024518 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:22 crc kubenswrapper[4945]: E1008 15:15:22.024571 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:22 crc kubenswrapper[4945]: E1008 15:15:22.250356 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 15:15:24 crc kubenswrapper[4945]: I1008 15:15:24.023532 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:24 crc kubenswrapper[4945]: I1008 15:15:24.023621 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:24 crc kubenswrapper[4945]: E1008 15:15:24.023676 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:24 crc kubenswrapper[4945]: I1008 15:15:24.023621 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:24 crc kubenswrapper[4945]: E1008 15:15:24.023859 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:24 crc kubenswrapper[4945]: I1008 15:15:24.023892 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:24 crc kubenswrapper[4945]: E1008 15:15:24.023899 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:24 crc kubenswrapper[4945]: E1008 15:15:24.024043 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:25 crc kubenswrapper[4945]: I1008 15:15:25.024549 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:15:25 crc kubenswrapper[4945]: E1008 15:15:25.024805 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9g2vg_openshift-ovn-kubernetes(28239584-598a-49d2-a9b0-189e4f4ad733)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" Oct 08 15:15:26 crc kubenswrapper[4945]: I1008 15:15:26.024222 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:26 crc kubenswrapper[4945]: I1008 15:15:26.024331 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:26 crc kubenswrapper[4945]: E1008 15:15:26.024349 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:26 crc kubenswrapper[4945]: I1008 15:15:26.024596 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:26 crc kubenswrapper[4945]: E1008 15:15:26.024599 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:26 crc kubenswrapper[4945]: E1008 15:15:26.024646 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:26 crc kubenswrapper[4945]: I1008 15:15:26.024909 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:26 crc kubenswrapper[4945]: E1008 15:15:26.025000 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:27 crc kubenswrapper[4945]: E1008 15:15:27.251895 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 15:15:28 crc kubenswrapper[4945]: I1008 15:15:28.023758 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:28 crc kubenswrapper[4945]: I1008 15:15:28.023841 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:28 crc kubenswrapper[4945]: E1008 15:15:28.023877 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:28 crc kubenswrapper[4945]: I1008 15:15:28.023902 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:28 crc kubenswrapper[4945]: I1008 15:15:28.023965 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:28 crc kubenswrapper[4945]: E1008 15:15:28.024022 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:28 crc kubenswrapper[4945]: E1008 15:15:28.024058 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:28 crc kubenswrapper[4945]: E1008 15:15:28.024148 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:30 crc kubenswrapper[4945]: I1008 15:15:30.023264 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:30 crc kubenswrapper[4945]: E1008 15:15:30.024043 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:30 crc kubenswrapper[4945]: I1008 15:15:30.023432 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:30 crc kubenswrapper[4945]: E1008 15:15:30.024480 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:30 crc kubenswrapper[4945]: I1008 15:15:30.023353 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:30 crc kubenswrapper[4945]: E1008 15:15:30.024737 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:30 crc kubenswrapper[4945]: I1008 15:15:30.023442 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:30 crc kubenswrapper[4945]: E1008 15:15:30.025006 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:31 crc kubenswrapper[4945]: I1008 15:15:31.023950 4945 scope.go:117] "RemoveContainer" containerID="892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203" Oct 08 15:15:31 crc kubenswrapper[4945]: I1008 15:15:31.718866 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/1.log" Oct 08 15:15:31 crc kubenswrapper[4945]: I1008 15:15:31.718943 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerStarted","Data":"e4e3fedefec003a21b24bf0cb252dbec6df8b7b9ba1b1d3af62aadd0be7e8bcd"} Oct 08 15:15:32 crc kubenswrapper[4945]: I1008 15:15:32.023299 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:32 crc kubenswrapper[4945]: E1008 15:15:32.024507 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:32 crc kubenswrapper[4945]: I1008 15:15:32.024543 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:32 crc kubenswrapper[4945]: I1008 15:15:32.024593 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:32 crc kubenswrapper[4945]: I1008 15:15:32.024556 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:32 crc kubenswrapper[4945]: E1008 15:15:32.024657 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:32 crc kubenswrapper[4945]: E1008 15:15:32.024704 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:32 crc kubenswrapper[4945]: E1008 15:15:32.024775 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:32 crc kubenswrapper[4945]: E1008 15:15:32.253350 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 15:15:34 crc kubenswrapper[4945]: I1008 15:15:34.023982 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:34 crc kubenswrapper[4945]: I1008 15:15:34.024035 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:34 crc kubenswrapper[4945]: I1008 15:15:34.023990 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:34 crc kubenswrapper[4945]: E1008 15:15:34.024221 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:34 crc kubenswrapper[4945]: I1008 15:15:34.024239 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:34 crc kubenswrapper[4945]: E1008 15:15:34.024379 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:34 crc kubenswrapper[4945]: E1008 15:15:34.024638 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:34 crc kubenswrapper[4945]: E1008 15:15:34.024741 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.023560 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.023713 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.024032 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:36 crc kubenswrapper[4945]: E1008 15:15:36.024225 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.024383 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.024550 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:36 crc kubenswrapper[4945]: E1008 15:15:36.024831 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:36 crc kubenswrapper[4945]: E1008 15:15:36.024657 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:36 crc kubenswrapper[4945]: E1008 15:15:36.024960 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.738545 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/3.log" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.742878 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerStarted","Data":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.743497 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.786352 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podStartSLOduration=116.786334144 podStartE2EDuration="1m56.786334144s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:36.784733413 +0000 UTC m=+146.138648334" watchObservedRunningTime="2025-10-08 15:15:36.786334144 +0000 UTC m=+146.140249045" Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.990195 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zj7s7"] Oct 08 15:15:36 crc kubenswrapper[4945]: I1008 15:15:36.990330 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:36 crc kubenswrapper[4945]: E1008 15:15:36.990438 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:37 crc kubenswrapper[4945]: E1008 15:15:37.255361 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 15:15:38 crc kubenswrapper[4945]: I1008 15:15:38.023284 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:38 crc kubenswrapper[4945]: I1008 15:15:38.023294 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:38 crc kubenswrapper[4945]: E1008 15:15:38.023426 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:38 crc kubenswrapper[4945]: E1008 15:15:38.023514 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:38 crc kubenswrapper[4945]: I1008 15:15:38.023627 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:38 crc kubenswrapper[4945]: E1008 15:15:38.023676 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:39 crc kubenswrapper[4945]: I1008 15:15:39.024209 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:39 crc kubenswrapper[4945]: E1008 15:15:39.024364 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:40 crc kubenswrapper[4945]: I1008 15:15:40.024221 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:40 crc kubenswrapper[4945]: I1008 15:15:40.024295 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:40 crc kubenswrapper[4945]: E1008 15:15:40.024364 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:40 crc kubenswrapper[4945]: I1008 15:15:40.024842 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:40 crc kubenswrapper[4945]: E1008 15:15:40.024959 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:40 crc kubenswrapper[4945]: E1008 15:15:40.024762 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:41 crc kubenswrapper[4945]: I1008 15:15:41.023907 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.024060 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zj7s7" podUID="6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0" Oct 08 15:15:41 crc kubenswrapper[4945]: I1008 15:15:41.817440 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.817685 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:17:43.817643262 +0000 UTC m=+273.171558203 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:41 crc kubenswrapper[4945]: I1008 15:15:41.817945 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.818168 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.818291 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:17:43.818271127 +0000 UTC m=+273.172186068 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: I1008 15:15:41.918859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:41 crc kubenswrapper[4945]: I1008 15:15:41.918935 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:41 crc kubenswrapper[4945]: I1008 15:15:41.919003 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919180 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919239 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919255 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919304 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919310 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 15:17:43.919284342 +0000 UTC m=+273.273199253 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919326 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919268 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919407 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 15:17:43.919381755 +0000 UTC m=+273.273296696 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919430 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:15:41 crc kubenswrapper[4945]: E1008 15:15:41.919514 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 15:17:43.919491657 +0000 UTC m=+273.273406588 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 15:15:42 crc kubenswrapper[4945]: I1008 15:15:42.023242 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:42 crc kubenswrapper[4945]: I1008 15:15:42.023296 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:42 crc kubenswrapper[4945]: E1008 15:15:42.023384 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 15:15:42 crc kubenswrapper[4945]: I1008 15:15:42.023260 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:42 crc kubenswrapper[4945]: E1008 15:15:42.023642 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 15:15:42 crc kubenswrapper[4945]: E1008 15:15:42.023635 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.024036 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.026786 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.026881 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.480409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.514916 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xghvq"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.515238 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sbkjb"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.515609 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.515918 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.517393 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.517636 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.517830 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.518043 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.518640 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j72zk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.519035 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.519915 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-95k2h"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.520238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.522551 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.523155 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.523491 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.523781 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.523802 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.523965 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.524171 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.524291 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.524389 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.524489 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.524593 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.524658 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.524713 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.528299 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.528508 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.528718 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.529643 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.529878 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.530095 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.530980 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.531230 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.531429 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.531609 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.531684 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.532067 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.546542 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550319 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550395 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-config\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550416 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sj54\" (UniqueName: \"kubernetes.io/projected/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-kube-api-access-2sj54\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550436 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-serving-cert\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550459 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-serving-cert\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550482 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khthc\" (UniqueName: \"kubernetes.io/projected/ec002292-f388-497c-b114-fb7818730f4d-kube-api-access-khthc\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550513 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-config\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550539 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.550559 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-client-ca\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.561913 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563185 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec002292-f388-497c-b114-fb7818730f4d-config\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563234 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-machine-approver-tls\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563266 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e75c21c-179f-4c78-b1f4-dab40ee201fb-audit-dir\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563289 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec002292-f388-497c-b114-fb7818730f4d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-etcd-serving-ca\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec002292-f388-497c-b114-fb7818730f4d-images\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563381 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-config\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563400 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f113bd-0207-40ce-9793-c7b0c901b885-serving-cert\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563432 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-config\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563455 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e75c21c-179f-4c78-b1f4-dab40ee201fb-node-pullsecrets\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-encryption-config\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563498 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-serving-cert\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563512 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-audit\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563531 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-etcd-client\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvn2n\" (UniqueName: \"kubernetes.io/projected/b5f113bd-0207-40ce-9793-c7b0c901b885-kube-api-access-cvn2n\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563589 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-client-ca\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563603 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krlz9\" (UniqueName: \"kubernetes.io/projected/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-kube-api-access-krlz9\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563628 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-image-import-ca\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjkb6\" (UniqueName: \"kubernetes.io/projected/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-kube-api-access-mjkb6\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-service-ca-bundle\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563676 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-auth-proxy-config\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563697 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spvhc\" (UniqueName: \"kubernetes.io/projected/8e75c21c-179f-4c78-b1f4-dab40ee201fb-kube-api-access-spvhc\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563711 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.563728 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-config\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.564781 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.566415 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b5mwn"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.566635 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.567074 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.567325 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.567499 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.568526 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.569296 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.569842 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.570304 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.570418 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.570482 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.570637 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.571237 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.571332 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.571489 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.571510 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.571757 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.572220 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kr5ql"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.572490 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.572759 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.574090 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.576471 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-d8prt"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.574875 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.576556 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.578517 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-hb8lk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.578747 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.578965 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.579238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.579264 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.579475 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.579238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.579905 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xghvq"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.580630 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-m7d4b"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.581292 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j72zk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.581346 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.582145 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.582496 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.583339 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.583995 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.584043 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.584293 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.584335 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.585095 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.585441 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.585759 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.590810 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.590875 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-922gb"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.591669 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.594166 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.595417 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.596508 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.596790 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.596934 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.599940 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.600092 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.600272 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.600294 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.600450 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.600478 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.614888 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.614937 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.615223 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.615399 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.615753 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.615758 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.615797 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.616125 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.616364 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.616545 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.616592 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.616669 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.615757 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.617443 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.618737 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.620699 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dl6kq"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.621305 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.623205 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.623284 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.623451 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.623715 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.623785 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.623909 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624014 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624034 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624103 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624208 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624316 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624458 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624540 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624707 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624889 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624472 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.625490 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sbkjb"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.627937 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.629153 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.629733 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624914 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.630296 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624208 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624504 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.624018 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.625293 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.625319 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.625388 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.625571 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.626645 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.627701 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.632598 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.632724 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.633028 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.633370 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-pr9cp"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.633859 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.634247 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.635601 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.636087 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.637085 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.638447 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.642426 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.646268 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.648035 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9r6ts"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.648660 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.649984 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.650428 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.650717 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.651082 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.653136 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.657500 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.658162 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.658751 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.659625 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665262 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665373 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec002292-f388-497c-b114-fb7818730f4d-config\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-machine-approver-tls\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665440 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b6a0a68-fe4e-4f50-af67-6d96384b1064-serving-cert\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665464 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-trusted-ca-bundle\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665487 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665507 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e75c21c-179f-4c78-b1f4-dab40ee201fb-audit-dir\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665529 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec002292-f388-497c-b114-fb7818730f4d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665562 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665586 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-etcd-serving-ca\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665604 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/805c4266-7d9c-4a26-af0d-9d25ca406fb8-proxy-tls\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665626 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-encryption-config\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9ggp\" (UniqueName: \"kubernetes.io/projected/9b6a0a68-fe4e-4f50-af67-6d96384b1064-kube-api-access-s9ggp\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665679 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665699 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-oauth-serving-cert\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665722 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd592\" (UniqueName: \"kubernetes.io/projected/8ef73f3d-3f61-499c-8806-cf041851986c-kube-api-access-xd592\") pod \"cluster-samples-operator-665b6dd947-c2cpk\" (UID: \"8ef73f3d-3f61-499c-8806-cf041851986c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665741 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l45wb\" (UniqueName: \"kubernetes.io/projected/44b02303-2b9b-448e-86e5-ed9bc3a59eec-kube-api-access-l45wb\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-config\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665781 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec002292-f388-497c-b114-fb7818730f4d-images\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665805 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-dir\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665832 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-oauth-config\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665851 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdac1f7c-1f67-43f8-8801-4014a03f9a14-config\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665874 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f113bd-0207-40ce-9793-c7b0c901b885-serving-cert\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665897 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665924 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665947 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-config\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665964 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-etcd-client\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.665984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e75c21c-179f-4c78-b1f4-dab40ee201fb-node-pullsecrets\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666003 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b6a0a68-fe4e-4f50-af67-6d96384b1064-trusted-ca\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdac1f7c-1f67-43f8-8801-4014a03f9a14-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cdac1f7c-1f67-43f8-8801-4014a03f9a14-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666199 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-encryption-config\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666219 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666269 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzhtq\" (UniqueName: \"kubernetes.io/projected/536ab60e-f81d-4b67-9528-fe765502f8e0-kube-api-access-xzhtq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666290 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-serving-cert\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666390 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666455 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-audit\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666477 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-etcd-client\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666498 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvn2n\" (UniqueName: \"kubernetes.io/projected/b5f113bd-0207-40ce-9793-c7b0c901b885-kube-api-access-cvn2n\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666524 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/536ab60e-f81d-4b67-9528-fe765502f8e0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-client-ca\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666792 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krlz9\" (UniqueName: \"kubernetes.io/projected/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-kube-api-access-krlz9\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666881 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-image-import-ca\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.666907 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjkb6\" (UniqueName: \"kubernetes.io/projected/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-kube-api-access-mjkb6\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667065 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667183 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/79d1c40a-ac76-4f69-94da-b701c1804a54-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667209 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-service-ca-bundle\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-auth-proxy-config\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667366 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-config\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzmhx\" (UniqueName: \"kubernetes.io/projected/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-kube-api-access-nzmhx\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667526 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-serving-cert\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667551 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spvhc\" (UniqueName: \"kubernetes.io/projected/8e75c21c-179f-4c78-b1f4-dab40ee201fb-kube-api-access-spvhc\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667637 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667657 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvfrf\" (UniqueName: \"kubernetes.io/projected/54eb5003-d347-47aa-b14a-391d1bd74bdb-kube-api-access-lvfrf\") pod \"downloads-7954f5f757-hb8lk\" (UID: \"54eb5003-d347-47aa-b14a-391d1bd74bdb\") " pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667681 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667744 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-service-ca\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667769 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-config\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667787 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79d1c40a-ac76-4f69-94da-b701c1804a54-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667810 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ef73f3d-3f61-499c-8806-cf041851986c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c2cpk\" (UID: \"8ef73f3d-3f61-499c-8806-cf041851986c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667907 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/79d1c40a-ac76-4f69-94da-b701c1804a54-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667938 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667959 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.667979 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668002 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc4qf\" (UniqueName: \"kubernetes.io/projected/8db81fa8-6d99-4529-a1a8-c573752dfb18-kube-api-access-sc4qf\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668034 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-serving-cert\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-config\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668094 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sj54\" (UniqueName: \"kubernetes.io/projected/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-kube-api-access-2sj54\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668132 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-serving-cert\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668239 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz2t5\" (UniqueName: \"kubernetes.io/projected/79d1c40a-ac76-4f69-94da-b701c1804a54-kube-api-access-tz2t5\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668281 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-serving-cert\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668317 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khthc\" (UniqueName: \"kubernetes.io/projected/ec002292-f388-497c-b114-fb7818730f4d-kube-api-access-khthc\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668323 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec002292-f388-497c-b114-fb7818730f4d-config\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668339 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbqnm\" (UniqueName: \"kubernetes.io/projected/805c4266-7d9c-4a26-af0d-9d25ca406fb8-kube-api-access-vbqnm\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668421 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-audit-policies\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668454 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44b02303-2b9b-448e-86e5-ed9bc3a59eec-audit-dir\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668490 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a0a68-fe4e-4f50-af67-6d96384b1064-config\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668528 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-config\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668565 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-client-ca\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/805c4266-7d9c-4a26-af0d-9d25ca406fb8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668682 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-policies\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.668777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536ab60e-f81d-4b67-9528-fe765502f8e0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.669473 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.669743 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e75c21c-179f-4c78-b1f4-dab40ee201fb-audit-dir\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.669780 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-config\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.673960 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e75c21c-179f-4c78-b1f4-dab40ee201fb-node-pullsecrets\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.674066 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.675207 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-etcd-serving-ca\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.676039 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-config\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.676061 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-client-ca\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.676233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-audit\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.676246 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-service-ca-bundle\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.676299 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-auth-proxy-config\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.677455 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.677493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-client-ca\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.677797 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.678061 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-config\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.678475 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec002292-f388-497c-b114-fb7818730f4d-images\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.679012 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.679104 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f113bd-0207-40ce-9793-c7b0c901b885-serving-cert\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.679583 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-config\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.679772 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.680315 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec002292-f388-497c-b114-fb7818730f4d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.681539 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-serving-cert\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.682070 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-serving-cert\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.682505 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-machine-approver-tls\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.683003 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-etcd-client\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.683681 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-serving-cert\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.685886 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e75c21c-179f-4c78-b1f4-dab40ee201fb-encryption-config\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.686332 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8e75c21c-179f-4c78-b1f4-dab40ee201fb-image-import-ca\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.687639 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-62xzf"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.688050 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.688461 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.688906 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m5fwt"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.689371 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tnlpm"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.689721 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6trg2"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.690536 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.690872 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.690891 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b5mwn"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.690903 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kr5ql"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.690915 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-922gb"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.690992 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.691234 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.691445 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.691638 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.691655 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.691657 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.691647 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.692020 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.692628 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-95k2h"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.695621 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-d8prt"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.695685 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hb8lk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.695748 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.696248 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-config\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.699698 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.699725 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-m7d4b"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.699736 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.700779 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.700948 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.701895 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.702863 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dl6kq"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.703945 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.704905 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.705924 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.708099 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m5fwt"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.708889 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9r6ts"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.709731 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.710732 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.711785 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.712842 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.715138 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.715203 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.716226 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.717264 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6trg2"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.717698 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.718515 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-62xzf"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.719853 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.723354 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tnlpm"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.725091 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.726588 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vkzx6"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.727525 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.728744 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2r4nc"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.729935 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.732245 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2r4nc"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.733859 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vkzx6"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.739457 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.740242 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7k5pl"] Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.741337 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.757325 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.769867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/79d1c40a-ac76-4f69-94da-b701c1804a54-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.769934 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.769956 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzmhx\" (UniqueName: \"kubernetes.io/projected/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-kube-api-access-nzmhx\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.769973 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-serving-cert\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.770016 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-config\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.770056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.770932 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771010 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-config\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771074 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771095 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvfrf\" (UniqueName: \"kubernetes.io/projected/54eb5003-d347-47aa-b14a-391d1bd74bdb-kube-api-access-lvfrf\") pod \"downloads-7954f5f757-hb8lk\" (UID: \"54eb5003-d347-47aa-b14a-391d1bd74bdb\") " pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771133 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-service-ca\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79d1c40a-ac76-4f69-94da-b701c1804a54-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771195 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ef73f3d-3f61-499c-8806-cf041851986c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c2cpk\" (UID: \"8ef73f3d-3f61-499c-8806-cf041851986c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771211 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/79d1c40a-ac76-4f69-94da-b701c1804a54-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771228 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771255 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771270 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc4qf\" (UniqueName: \"kubernetes.io/projected/8db81fa8-6d99-4529-a1a8-c573752dfb18-kube-api-access-sc4qf\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771285 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-serving-cert\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771318 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz2t5\" (UniqueName: \"kubernetes.io/projected/79d1c40a-ac76-4f69-94da-b701c1804a54-kube-api-access-tz2t5\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771336 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbqnm\" (UniqueName: \"kubernetes.io/projected/805c4266-7d9c-4a26-af0d-9d25ca406fb8-kube-api-access-vbqnm\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771353 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-audit-policies\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771368 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44b02303-2b9b-448e-86e5-ed9bc3a59eec-audit-dir\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771389 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a0a68-fe4e-4f50-af67-6d96384b1064-config\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771440 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-policies\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/805c4266-7d9c-4a26-af0d-9d25ca406fb8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771477 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536ab60e-f81d-4b67-9528-fe765502f8e0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771491 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-trusted-ca-bundle\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771507 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771533 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b6a0a68-fe4e-4f50-af67-6d96384b1064-serving-cert\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771548 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771565 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/805c4266-7d9c-4a26-af0d-9d25ca406fb8-proxy-tls\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771583 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-encryption-config\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771611 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9ggp\" (UniqueName: \"kubernetes.io/projected/9b6a0a68-fe4e-4f50-af67-6d96384b1064-kube-api-access-s9ggp\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771627 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771643 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-oauth-serving-cert\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771659 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd592\" (UniqueName: \"kubernetes.io/projected/8ef73f3d-3f61-499c-8806-cf041851986c-kube-api-access-xd592\") pod \"cluster-samples-operator-665b6dd947-c2cpk\" (UID: \"8ef73f3d-3f61-499c-8806-cf041851986c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771676 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l45wb\" (UniqueName: \"kubernetes.io/projected/44b02303-2b9b-448e-86e5-ed9bc3a59eec-kube-api-access-l45wb\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771692 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-dir\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-oauth-config\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdac1f7c-1f67-43f8-8801-4014a03f9a14-config\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771742 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771781 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-etcd-client\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771797 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b6a0a68-fe4e-4f50-af67-6d96384b1064-trusted-ca\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771812 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdac1f7c-1f67-43f8-8801-4014a03f9a14-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771831 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cdac1f7c-1f67-43f8-8801-4014a03f9a14-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771846 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzhtq\" (UniqueName: \"kubernetes.io/projected/536ab60e-f81d-4b67-9528-fe765502f8e0-kube-api-access-xzhtq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771865 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771883 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.771910 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/536ab60e-f81d-4b67-9528-fe765502f8e0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.772007 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.772865 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.773217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b6a0a68-fe4e-4f50-af67-6d96384b1064-config\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.773218 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536ab60e-f81d-4b67-9528-fe765502f8e0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.773741 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-policies\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.773860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44b02303-2b9b-448e-86e5-ed9bc3a59eec-audit-dir\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.773980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-audit-policies\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.774009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/805c4266-7d9c-4a26-af0d-9d25ca406fb8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.774941 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b6a0a68-fe4e-4f50-af67-6d96384b1064-trusted-ca\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.775201 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79d1c40a-ac76-4f69-94da-b701c1804a54-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.775840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdac1f7c-1f67-43f8-8801-4014a03f9a14-config\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.775861 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ef73f3d-3f61-499c-8806-cf041851986c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c2cpk\" (UID: \"8ef73f3d-3f61-499c-8806-cf041851986c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.775861 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-serving-cert\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.776010 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b6a0a68-fe4e-4f50-af67-6d96384b1064-serving-cert\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.776865 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/536ab60e-f81d-4b67-9528-fe765502f8e0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.776922 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.777354 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-serving-cert\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.777451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44b02303-2b9b-448e-86e5-ed9bc3a59eec-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.778352 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-oauth-serving-cert\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.778390 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.778651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-dir\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.778661 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.778771 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.778789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-etcd-client\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.779242 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.779475 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-trusted-ca-bundle\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.779490 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.779851 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.780278 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/44b02303-2b9b-448e-86e5-ed9bc3a59eec-encryption-config\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.780384 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.781064 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/79d1c40a-ac76-4f69-94da-b701c1804a54-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.781607 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdac1f7c-1f67-43f8-8801-4014a03f9a14-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.781655 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-oauth-config\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.781977 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.782048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.784501 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-service-ca\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.798209 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.824189 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.838319 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.857866 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.877623 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.898082 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.917668 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.937943 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.957919 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.977774 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 15:15:43 crc kubenswrapper[4945]: I1008 15:15:43.998359 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.017747 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.023434 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.023633 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.023718 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.037668 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.058006 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.079566 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.098698 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.119475 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.137838 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.157960 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.178365 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.199225 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.238225 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.251343 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/805c4266-7d9c-4a26-af0d-9d25ca406fb8-proxy-tls\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.258507 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.278596 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.297939 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.320182 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.339234 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.358427 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.377616 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.398246 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.438853 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.459194 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.478837 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.498207 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.518662 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.537370 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.559226 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.579003 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.598212 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.618765 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.638881 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.656971 4945 request.go:700] Waited for 1.005639242s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.658922 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.678628 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.698427 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.718191 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.739414 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.758555 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.779402 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.798842 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.833867 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krlz9\" (UniqueName: \"kubernetes.io/projected/acf1d244-e7ed-41ec-8bc4-ef88aedb1411-kube-api-access-krlz9\") pod \"machine-approver-56656f9798-5k2tz\" (UID: \"acf1d244-e7ed-41ec-8bc4-ef88aedb1411\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.838184 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.849816 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.860939 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 15:15:44 crc kubenswrapper[4945]: W1008 15:15:44.866750 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacf1d244_e7ed_41ec_8bc4_ef88aedb1411.slice/crio-eb55d1bdf6fd3b88781fa9d3bf8db5fb476daf4456070961a46f3c55df000cfb WatchSource:0}: Error finding container eb55d1bdf6fd3b88781fa9d3bf8db5fb476daf4456070961a46f3c55df000cfb: Status 404 returned error can't find the container with id eb55d1bdf6fd3b88781fa9d3bf8db5fb476daf4456070961a46f3c55df000cfb Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.878050 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.916202 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spvhc\" (UniqueName: \"kubernetes.io/projected/8e75c21c-179f-4c78-b1f4-dab40ee201fb-kube-api-access-spvhc\") pod \"apiserver-76f77b778f-sbkjb\" (UID: \"8e75c21c-179f-4c78-b1f4-dab40ee201fb\") " pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.933498 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjkb6\" (UniqueName: \"kubernetes.io/projected/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-kube-api-access-mjkb6\") pod \"route-controller-manager-6576b87f9c-477zg\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.951952 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sj54\" (UniqueName: \"kubernetes.io/projected/075d448b-cf3c-47e4-a9bb-92d9a8295a2d-kube-api-access-2sj54\") pod \"authentication-operator-69f744f599-95k2h\" (UID: \"075d448b-cf3c-47e4-a9bb-92d9a8295a2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.972848 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khthc\" (UniqueName: \"kubernetes.io/projected/ec002292-f388-497c-b114-fb7818730f4d-kube-api-access-khthc\") pod \"machine-api-operator-5694c8668f-j72zk\" (UID: \"ec002292-f388-497c-b114-fb7818730f4d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.993783 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvn2n\" (UniqueName: \"kubernetes.io/projected/b5f113bd-0207-40ce-9793-c7b0c901b885-kube-api-access-cvn2n\") pod \"controller-manager-879f6c89f-xghvq\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:44 crc kubenswrapper[4945]: I1008 15:15:44.997879 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.018981 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.038554 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.058391 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.079210 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.081963 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.092266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.099939 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.118478 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.138420 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.141261 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.158646 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.158999 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.168001 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.179223 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.198828 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.218622 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.247947 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.258065 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.278803 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.300626 4945 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.321851 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.338520 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.347770 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sbkjb"] Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.359128 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.377974 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.397656 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.398841 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xghvq"] Oct 08 15:15:45 crc kubenswrapper[4945]: W1008 15:15:45.399294 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e75c21c_179f_4c78_b1f4_dab40ee201fb.slice/crio-1b0c8c0f32d90910bb925c5866ec0092ad63733550e47afd585ceb836f8ab726 WatchSource:0}: Error finding container 1b0c8c0f32d90910bb925c5866ec0092ad63733550e47afd585ceb836f8ab726: Status 404 returned error can't find the container with id 1b0c8c0f32d90910bb925c5866ec0092ad63733550e47afd585ceb836f8ab726 Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.405240 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j72zk"] Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.408382 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg"] Oct 08 15:15:45 crc kubenswrapper[4945]: W1008 15:15:45.413334 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec002292_f388_497c_b114_fb7818730f4d.slice/crio-1dee636b8c9104bcaca190e910abe4073d500cd3deb739529bd348747d71f7cb WatchSource:0}: Error finding container 1dee636b8c9104bcaca190e910abe4073d500cd3deb739529bd348747d71f7cb: Status 404 returned error can't find the container with id 1dee636b8c9104bcaca190e910abe4073d500cd3deb739529bd348747d71f7cb Oct 08 15:15:45 crc kubenswrapper[4945]: W1008 15:15:45.415222 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5f113bd_0207_40ce_9793_c7b0c901b885.slice/crio-48a002ba0ed1591f492bfdbfe64365da649bb12f1789f7247b952f391e33811e WatchSource:0}: Error finding container 48a002ba0ed1591f492bfdbfe64365da649bb12f1789f7247b952f391e33811e: Status 404 returned error can't find the container with id 48a002ba0ed1591f492bfdbfe64365da649bb12f1789f7247b952f391e33811e Oct 08 15:15:45 crc kubenswrapper[4945]: W1008 15:15:45.416769 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod500dbefa_9b5b_4dcf_8cc2_a70b4fa35193.slice/crio-bd9bf508c258f0143c49eb8de9769c51d86f272e2a4ef694e59113d4ab8aebea WatchSource:0}: Error finding container bd9bf508c258f0143c49eb8de9769c51d86f272e2a4ef694e59113d4ab8aebea: Status 404 returned error can't find the container with id bd9bf508c258f0143c49eb8de9769c51d86f272e2a4ef694e59113d4ab8aebea Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.418463 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.432616 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-95k2h"] Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.438340 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 15:15:45 crc kubenswrapper[4945]: W1008 15:15:45.456727 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod075d448b_cf3c_47e4_a9bb_92d9a8295a2d.slice/crio-7adcbae30b464e5164d37cfabda4e7583dfeebac39d92163c870fd8ade32c53b WatchSource:0}: Error finding container 7adcbae30b464e5164d37cfabda4e7583dfeebac39d92163c870fd8ade32c53b: Status 404 returned error can't find the container with id 7adcbae30b464e5164d37cfabda4e7583dfeebac39d92163c870fd8ade32c53b Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.458171 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.477751 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.498552 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.517801 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.538453 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.558180 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.579631 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.598747 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.618368 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.638342 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.658057 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.676484 4945 request.go:700] Waited for 1.934547746s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.678168 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.715912 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/79d1c40a-ac76-4f69-94da-b701c1804a54-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.737037 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzmhx\" (UniqueName: \"kubernetes.io/projected/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-kube-api-access-nzmhx\") pod \"console-f9d7485db-kr5ql\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.753219 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz2t5\" (UniqueName: \"kubernetes.io/projected/79d1c40a-ac76-4f69-94da-b701c1804a54-kube-api-access-tz2t5\") pod \"cluster-image-registry-operator-dc59b4c8b-xlgt2\" (UID: \"79d1c40a-ac76-4f69-94da-b701c1804a54\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.772158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbqnm\" (UniqueName: \"kubernetes.io/projected/805c4266-7d9c-4a26-af0d-9d25ca406fb8-kube-api-access-vbqnm\") pod \"machine-config-controller-84d6567774-zjs2d\" (UID: \"805c4266-7d9c-4a26-af0d-9d25ca406fb8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.773305 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" event={"ID":"b5f113bd-0207-40ce-9793-c7b0c901b885","Type":"ContainerStarted","Data":"48a002ba0ed1591f492bfdbfe64365da649bb12f1789f7247b952f391e33811e"} Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.782566 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" event={"ID":"075d448b-cf3c-47e4-a9bb-92d9a8295a2d","Type":"ContainerStarted","Data":"7adcbae30b464e5164d37cfabda4e7583dfeebac39d92163c870fd8ade32c53b"} Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.784549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" event={"ID":"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193","Type":"ContainerStarted","Data":"bd9bf508c258f0143c49eb8de9769c51d86f272e2a4ef694e59113d4ab8aebea"} Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.786224 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" event={"ID":"acf1d244-e7ed-41ec-8bc4-ef88aedb1411","Type":"ContainerStarted","Data":"015f85069cec53baa026c85a4b303c51d8f6e280bac6003a5001011bb1c94264"} Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.786273 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" event={"ID":"acf1d244-e7ed-41ec-8bc4-ef88aedb1411","Type":"ContainerStarted","Data":"eb55d1bdf6fd3b88781fa9d3bf8db5fb476daf4456070961a46f3c55df000cfb"} Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.788761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" event={"ID":"ec002292-f388-497c-b114-fb7818730f4d","Type":"ContainerStarted","Data":"1dee636b8c9104bcaca190e910abe4073d500cd3deb739529bd348747d71f7cb"} Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.791461 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" event={"ID":"8e75c21c-179f-4c78-b1f4-dab40ee201fb","Type":"ContainerStarted","Data":"1b0c8c0f32d90910bb925c5866ec0092ad63733550e47afd585ceb836f8ab726"} Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.800895 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvfrf\" (UniqueName: \"kubernetes.io/projected/54eb5003-d347-47aa-b14a-391d1bd74bdb-kube-api-access-lvfrf\") pod \"downloads-7954f5f757-hb8lk\" (UID: \"54eb5003-d347-47aa-b14a-391d1bd74bdb\") " pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.814100 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cdac1f7c-1f67-43f8-8801-4014a03f9a14-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-429v8\" (UID: \"cdac1f7c-1f67-43f8-8801-4014a03f9a14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.831765 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.853478 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzhtq\" (UniqueName: \"kubernetes.io/projected/536ab60e-f81d-4b67-9528-fe765502f8e0-kube-api-access-xzhtq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nfdsl\" (UID: \"536ab60e-f81d-4b67-9528-fe765502f8e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.857437 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.864539 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc4qf\" (UniqueName: \"kubernetes.io/projected/8db81fa8-6d99-4529-a1a8-c573752dfb18-kube-api-access-sc4qf\") pod \"oauth-openshift-558db77b4-d8prt\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.866028 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.875347 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l45wb\" (UniqueName: \"kubernetes.io/projected/44b02303-2b9b-448e-86e5-ed9bc3a59eec-kube-api-access-l45wb\") pod \"apiserver-7bbb656c7d-pjg7s\" (UID: \"44b02303-2b9b-448e-86e5-ed9bc3a59eec\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.880819 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.900452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9ggp\" (UniqueName: \"kubernetes.io/projected/9b6a0a68-fe4e-4f50-af67-6d96384b1064-kube-api-access-s9ggp\") pod \"console-operator-58897d9998-b5mwn\" (UID: \"9b6a0a68-fe4e-4f50-af67-6d96384b1064\") " pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.920348 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.921336 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.924672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd592\" (UniqueName: \"kubernetes.io/projected/8ef73f3d-3f61-499c-8806-cf041851986c-kube-api-access-xd592\") pod \"cluster-samples-operator-665b6dd947-c2cpk\" (UID: \"8ef73f3d-3f61-499c-8806-cf041851986c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.928882 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.943797 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.959411 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.966458 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" Oct 08 15:15:45 crc kubenswrapper[4945]: I1008 15:15:45.985859 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103841 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c56509df-b9b8-4857-929f-2a2667dc9e58-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103871 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/483f2b9f-6fcd-4569-ab81-a6f265083617-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103897 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c56509df-b9b8-4857-929f-2a2667dc9e58-config\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dca32bb-bf0b-4632-8fe5-af683164904a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103931 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-serving-cert\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103948 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppfdh\" (UniqueName: \"kubernetes.io/projected/fd7496df-c90b-454b-b107-6345db7738ca-kube-api-access-ppfdh\") pod \"dns-operator-744455d44c-m7d4b\" (UID: \"fd7496df-c90b-454b-b107-6345db7738ca\") " pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103961 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-client\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103977 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5fec4a94-36ab-4892-a13c-35a21108d2ba-installation-pull-secrets\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.103992 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sprf\" (UniqueName: \"kubernetes.io/projected/483f2b9f-6fcd-4569-ab81-a6f265083617-kube-api-access-4sprf\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104007 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/556ada8d-ee27-46a9-94cd-a0ec52b077aa-metrics-tls\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104027 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpgq2\" (UniqueName: \"kubernetes.io/projected/f973b44d-6296-408b-9d14-31d56fdd1021-kube-api-access-hpgq2\") pod \"control-plane-machine-set-operator-78cbb6b69f-mn52j\" (UID: \"f973b44d-6296-408b-9d14-31d56fdd1021\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104047 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/556ada8d-ee27-46a9-94cd-a0ec52b077aa-bound-sa-token\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104122 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f973b44d-6296-408b-9d14-31d56fdd1021-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mn52j\" (UID: \"f973b44d-6296-408b-9d14-31d56fdd1021\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104142 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-certificates\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104166 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-bound-sa-token\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104183 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4ndj\" (UniqueName: \"kubernetes.io/projected/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-kube-api-access-v4ndj\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104198 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-config\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104215 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951bacb4-e545-4279-ad11-b791aed0aa1e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104248 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dca32bb-bf0b-4632-8fe5-af683164904a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104265 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c56509df-b9b8-4857-929f-2a2667dc9e58-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104283 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-trusted-ca\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104303 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104318 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhhn8\" (UniqueName: \"kubernetes.io/projected/951bacb4-e545-4279-ad11-b791aed0aa1e-kube-api-access-vhhn8\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104333 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-664pd\" (UniqueName: \"kubernetes.io/projected/556ada8d-ee27-46a9-94cd-a0ec52b077aa-kube-api-access-664pd\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104361 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-tls\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104410 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-service-ca\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951bacb4-e545-4279-ad11-b791aed0aa1e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/483f2b9f-6fcd-4569-ab81-a6f265083617-serving-cert\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104466 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5fec4a94-36ab-4892-a13c-35a21108d2ba-ca-trust-extracted\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104480 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2dca32bb-bf0b-4632-8fe5-af683164904a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104494 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd7496df-c90b-454b-b107-6345db7738ca-metrics-tls\") pod \"dns-operator-744455d44c-m7d4b\" (UID: \"fd7496df-c90b-454b-b107-6345db7738ca\") " pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb7bl\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-kube-api-access-wb7bl\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104523 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556ada8d-ee27-46a9-94cd-a0ec52b077aa-trusted-ca\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.104538 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-ca\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.104973 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:46.604956347 +0000 UTC m=+155.958871328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.108635 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.124611 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.153954 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.197031 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.201939 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-d8prt"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.205799 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206077 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0debf7-f7f7-4445-b5be-fa96774cbcf1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b6msk\" (UID: \"cc0debf7-f7f7-4445-b5be-fa96774cbcf1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.206143 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:46.706100705 +0000 UTC m=+156.060015606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206201 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-config\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206239 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mqsr\" (UniqueName: \"kubernetes.io/projected/87d27326-3c94-424c-af7e-6defb922382d-kube-api-access-9mqsr\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206288 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951bacb4-e545-4279-ad11-b791aed0aa1e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206306 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76a2fafe-8637-4067-a89b-cf6756ba3e09-srv-cert\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206321 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ad1a76d0-a604-4c89-921d-aed33f60fd01-webhook-cert\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206335 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-csi-data-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206350 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-plugins-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206409 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-registration-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206427 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc5p6\" (UniqueName: \"kubernetes.io/projected/fe723d30-456e-491e-84ed-141775f2592d-kube-api-access-wc5p6\") pod \"ingress-canary-2r4nc\" (UID: \"fe723d30-456e-491e-84ed-141775f2592d\") " pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dca32bb-bf0b-4632-8fe5-af683164904a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206471 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c56509df-b9b8-4857-929f-2a2667dc9e58-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206497 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-trusted-ca\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206523 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhhn8\" (UniqueName: \"kubernetes.io/projected/951bacb4-e545-4279-ad11-b791aed0aa1e-kube-api-access-vhhn8\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpxcz\" (UniqueName: \"kubernetes.io/projected/76a2fafe-8637-4067-a89b-cf6756ba3e09-kube-api-access-lpxcz\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206576 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206611 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-664pd\" (UniqueName: \"kubernetes.io/projected/556ada8d-ee27-46a9-94cd-a0ec52b077aa-kube-api-access-664pd\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206699 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-tls\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206718 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-service-ca\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206734 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951bacb4-e545-4279-ad11-b791aed0aa1e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206763 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/483f2b9f-6fcd-4569-ab81-a6f265083617-serving-cert\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206824 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-socket-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206856 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2dca32bb-bf0b-4632-8fe5-af683164904a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206877 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5fec4a94-36ab-4892-a13c-35a21108d2ba-ca-trust-extracted\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206893 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd7496df-c90b-454b-b107-6345db7738ca-metrics-tls\") pod \"dns-operator-744455d44c-m7d4b\" (UID: \"fd7496df-c90b-454b-b107-6345db7738ca\") " pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206920 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb7bl\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-kube-api-access-wb7bl\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556ada8d-ee27-46a9-94cd-a0ec52b077aa-trusted-ca\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206954 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b32adce4-0f8a-433e-8643-a9b35e02fd8d-service-ca-bundle\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.206990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-ca\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207007 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ndx2\" (UniqueName: \"kubernetes.io/projected/ad1a76d0-a604-4c89-921d-aed33f60fd01-kube-api-access-9ndx2\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207045 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ba793aa-5158-46a3-8ea6-1e91389524d1-secret-volume\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207085 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207117 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c56509df-b9b8-4857-929f-2a2667dc9e58-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207156 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-proxy-tls\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207189 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/483f2b9f-6fcd-4569-ab81-a6f265083617-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207227 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c56509df-b9b8-4857-929f-2a2667dc9e58-config\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207244 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dca32bb-bf0b-4632-8fe5-af683164904a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.207262 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbztx\" (UniqueName: \"kubernetes.io/projected/77e064ce-ee89-43e9-b600-a5ecf951cead-kube-api-access-hbztx\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208416 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-serving-cert\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208466 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppfdh\" (UniqueName: \"kubernetes.io/projected/fd7496df-c90b-454b-b107-6345db7738ca-kube-api-access-ppfdh\") pod \"dns-operator-744455d44c-m7d4b\" (UID: \"fd7496df-c90b-454b-b107-6345db7738ca\") " pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208488 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-client\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.208829 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:46.708814933 +0000 UTC m=+156.062729834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208864 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5fec4a94-36ab-4892-a13c-35a21108d2ba-installation-pull-secrets\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208896 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sprf\" (UniqueName: \"kubernetes.io/projected/483f2b9f-6fcd-4569-ab81-a6f265083617-kube-api-access-4sprf\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208922 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v59d\" (UniqueName: \"kubernetes.io/projected/17dcfe6a-c319-4ac4-82a9-407601bf589b-kube-api-access-5v59d\") pod \"multus-admission-controller-857f4d67dd-9r6ts\" (UID: \"17dcfe6a-c319-4ac4-82a9-407601bf589b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208940 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9afc51-40c4-46e2-8fdf-2d29f7520271-config\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/556ada8d-ee27-46a9-94cd-a0ec52b077aa-metrics-tls\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.209023 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ad1a76d0-a604-4c89-921d-aed33f60fd01-apiservice-cert\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.209039 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/87d27326-3c94-424c-af7e-6defb922382d-srv-cert\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.209055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw27x\" (UniqueName: \"kubernetes.io/projected/e76deaf4-7012-4889-87bf-1310e42a6f46-kube-api-access-tw27x\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.209074 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/dba538a8-b800-4492-b281-b4eb25873b71-certs\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.209069 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951bacb4-e545-4279-ad11-b791aed0aa1e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.209094 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-images\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.208830 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-service-ca\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.210647 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-config\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213519 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/483f2b9f-6fcd-4569-ab81-a6f265083617-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.211585 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-mountpoint-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556ada8d-ee27-46a9-94cd-a0ec52b077aa-trusted-ca\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213681 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/87d27326-3c94-424c-af7e-6defb922382d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213712 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpgq2\" (UniqueName: \"kubernetes.io/projected/f973b44d-6296-408b-9d14-31d56fdd1021-kube-api-access-hpgq2\") pod \"control-plane-machine-set-operator-78cbb6b69f-mn52j\" (UID: \"f973b44d-6296-408b-9d14-31d56fdd1021\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213736 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f9afc51-40c4-46e2-8fdf-2d29f7520271-serving-cert\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213759 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2z64\" (UniqueName: \"kubernetes.io/projected/86210512-22db-4bf6-876b-f10593c85e02-kube-api-access-b2z64\") pod \"migrator-59844c95c7-k8w22\" (UID: \"86210512-22db-4bf6-876b-f10593c85e02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213783 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64dlt\" (UniqueName: \"kubernetes.io/projected/f65f8689-77b0-4873-af9d-d82f5ec79c53-kube-api-access-64dlt\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213807 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-metrics-certs\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213934 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77e064ce-ee89-43e9-b600-a5ecf951cead-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.213990 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-ca\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214021 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h6j2\" (UniqueName: \"kubernetes.io/projected/0f9afc51-40c4-46e2-8fdf-2d29f7520271-kube-api-access-6h6j2\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214065 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ad1a76d0-a604-4c89-921d-aed33f60fd01-tmpfs\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214089 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/857289f7-30a2-42ed-8004-188b9e0a1626-signing-key\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6d6x\" (UniqueName: \"kubernetes.io/projected/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-kube-api-access-r6d6x\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214315 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/556ada8d-ee27-46a9-94cd-a0ec52b077aa-bound-sa-token\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214356 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-auth-proxy-config\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214381 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-default-certificate\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214471 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65f8689-77b0-4873-af9d-d82f5ec79c53-config-volume\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214514 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17dcfe6a-c319-4ac4-82a9-407601bf589b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9r6ts\" (UID: \"17dcfe6a-c319-4ac4-82a9-407601bf589b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76a2fafe-8637-4067-a89b-cf6756ba3e09-profile-collector-cert\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214561 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f65f8689-77b0-4873-af9d-d82f5ec79c53-metrics-tls\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214565 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c56509df-b9b8-4857-929f-2a2667dc9e58-config\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214882 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxxnw\" (UniqueName: \"kubernetes.io/projected/dba538a8-b800-4492-b281-b4eb25873b71-kube-api-access-bxxnw\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214949 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77e064ce-ee89-43e9-b600-a5ecf951cead-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.214981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f973b44d-6296-408b-9d14-31d56fdd1021-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mn52j\" (UID: \"f973b44d-6296-408b-9d14-31d56fdd1021\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.215005 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fe723d30-456e-491e-84ed-141775f2592d-cert\") pod \"ingress-canary-2r4nc\" (UID: \"fe723d30-456e-491e-84ed-141775f2592d\") " pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.215844 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5fec4a94-36ab-4892-a13c-35a21108d2ba-ca-trust-extracted\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.216045 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-trusted-ca\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.216371 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/483f2b9f-6fcd-4569-ab81-a6f265083617-serving-cert\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.216409 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c56509df-b9b8-4857-929f-2a2667dc9e58-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.216911 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-tls\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.216996 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-certificates\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.217704 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkkxp\" (UniqueName: \"kubernetes.io/projected/857289f7-30a2-42ed-8004-188b9e0a1626-kube-api-access-wkkxp\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.217780 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ft2\" (UniqueName: \"kubernetes.io/projected/2ba793aa-5158-46a3-8ea6-1e91389524d1-kube-api-access-j9ft2\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.217821 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.217902 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/dba538a8-b800-4492-b281-b4eb25873b71-node-bootstrap-token\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.217956 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-stats-auth\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.217978 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mfmk\" (UniqueName: \"kubernetes.io/projected/cc0debf7-f7f7-4445-b5be-fa96774cbcf1-kube-api-access-7mfmk\") pod \"package-server-manager-789f6589d5-b6msk\" (UID: \"cc0debf7-f7f7-4445-b5be-fa96774cbcf1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.218015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-bound-sa-token\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.218050 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkkxm\" (UniqueName: \"kubernetes.io/projected/b32adce4-0f8a-433e-8643-a9b35e02fd8d-kube-api-access-tkkxm\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.218096 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/857289f7-30a2-42ed-8004-188b9e0a1626-signing-cabundle\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.218145 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvjcd\" (UniqueName: \"kubernetes.io/projected/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-kube-api-access-rvjcd\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.218167 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ba793aa-5158-46a3-8ea6-1e91389524d1-config-volume\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.218196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4ndj\" (UniqueName: \"kubernetes.io/projected/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-kube-api-access-v4ndj\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.218277 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-certificates\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.220644 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5fec4a94-36ab-4892-a13c-35a21108d2ba-installation-pull-secrets\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.220980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-serving-cert\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.221688 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951bacb4-e545-4279-ad11-b791aed0aa1e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.221697 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd7496df-c90b-454b-b107-6345db7738ca-metrics-tls\") pod \"dns-operator-744455d44c-m7d4b\" (UID: \"fd7496df-c90b-454b-b107-6345db7738ca\") " pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.225625 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dca32bb-bf0b-4632-8fe5-af683164904a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.225747 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/556ada8d-ee27-46a9-94cd-a0ec52b077aa-metrics-tls\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.225838 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f973b44d-6296-408b-9d14-31d56fdd1021-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mn52j\" (UID: \"f973b44d-6296-408b-9d14-31d56fdd1021\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.226199 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-etcd-client\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.226713 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dca32bb-bf0b-4632-8fe5-af683164904a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.247724 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.271407 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2dca32bb-bf0b-4632-8fe5-af683164904a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zq9l5\" (UID: \"2dca32bb-bf0b-4632-8fe5-af683164904a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.276079 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.286962 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.298418 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb7bl\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-kube-api-access-wb7bl\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.299699 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhhn8\" (UniqueName: \"kubernetes.io/projected/951bacb4-e545-4279-ad11-b791aed0aa1e-kube-api-access-vhhn8\") pod \"openshift-apiserver-operator-796bbdcf4f-mdsqd\" (UID: \"951bacb4-e545-4279-ad11-b791aed0aa1e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.320514 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.321210 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:46.821186822 +0000 UTC m=+156.175101723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321271 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-proxy-tls\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321317 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbztx\" (UniqueName: \"kubernetes.io/projected/77e064ce-ee89-43e9-b600-a5ecf951cead-kube-api-access-hbztx\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321375 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v59d\" (UniqueName: \"kubernetes.io/projected/17dcfe6a-c319-4ac4-82a9-407601bf589b-kube-api-access-5v59d\") pod \"multus-admission-controller-857f4d67dd-9r6ts\" (UID: \"17dcfe6a-c319-4ac4-82a9-407601bf589b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321396 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9afc51-40c4-46e2-8fdf-2d29f7520271-config\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321424 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ad1a76d0-a604-4c89-921d-aed33f60fd01-apiservice-cert\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321447 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/87d27326-3c94-424c-af7e-6defb922382d-srv-cert\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321468 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw27x\" (UniqueName: \"kubernetes.io/projected/e76deaf4-7012-4889-87bf-1310e42a6f46-kube-api-access-tw27x\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321492 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/dba538a8-b800-4492-b281-b4eb25873b71-certs\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321530 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-mountpoint-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-images\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/87d27326-3c94-424c-af7e-6defb922382d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321609 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2z64\" (UniqueName: \"kubernetes.io/projected/86210512-22db-4bf6-876b-f10593c85e02-kube-api-access-b2z64\") pod \"migrator-59844c95c7-k8w22\" (UID: \"86210512-22db-4bf6-876b-f10593c85e02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64dlt\" (UniqueName: \"kubernetes.io/projected/f65f8689-77b0-4873-af9d-d82f5ec79c53-kube-api-access-64dlt\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321653 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-metrics-certs\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321679 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77e064ce-ee89-43e9-b600-a5ecf951cead-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321703 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f9afc51-40c4-46e2-8fdf-2d29f7520271-serving-cert\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321728 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h6j2\" (UniqueName: \"kubernetes.io/projected/0f9afc51-40c4-46e2-8fdf-2d29f7520271-kube-api-access-6h6j2\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321748 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ad1a76d0-a604-4c89-921d-aed33f60fd01-tmpfs\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/857289f7-30a2-42ed-8004-188b9e0a1626-signing-key\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321818 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6d6x\" (UniqueName: \"kubernetes.io/projected/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-kube-api-access-r6d6x\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-auth-proxy-config\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321883 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-default-certificate\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65f8689-77b0-4873-af9d-d82f5ec79c53-config-volume\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321957 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17dcfe6a-c319-4ac4-82a9-407601bf589b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9r6ts\" (UID: \"17dcfe6a-c319-4ac4-82a9-407601bf589b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.321981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76a2fafe-8637-4067-a89b-cf6756ba3e09-profile-collector-cert\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322003 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f65f8689-77b0-4873-af9d-d82f5ec79c53-metrics-tls\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322030 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxxnw\" (UniqueName: \"kubernetes.io/projected/dba538a8-b800-4492-b281-b4eb25873b71-kube-api-access-bxxnw\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322055 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77e064ce-ee89-43e9-b600-a5ecf951cead-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322078 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fe723d30-456e-491e-84ed-141775f2592d-cert\") pod \"ingress-canary-2r4nc\" (UID: \"fe723d30-456e-491e-84ed-141775f2592d\") " pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkkxp\" (UniqueName: \"kubernetes.io/projected/857289f7-30a2-42ed-8004-188b9e0a1626-kube-api-access-wkkxp\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322172 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ft2\" (UniqueName: \"kubernetes.io/projected/2ba793aa-5158-46a3-8ea6-1e91389524d1-kube-api-access-j9ft2\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322233 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/dba538a8-b800-4492-b281-b4eb25873b71-node-bootstrap-token\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322261 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mfmk\" (UniqueName: \"kubernetes.io/projected/cc0debf7-f7f7-4445-b5be-fa96774cbcf1-kube-api-access-7mfmk\") pod \"package-server-manager-789f6589d5-b6msk\" (UID: \"cc0debf7-f7f7-4445-b5be-fa96774cbcf1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322284 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-stats-auth\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322322 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkkxm\" (UniqueName: \"kubernetes.io/projected/b32adce4-0f8a-433e-8643-a9b35e02fd8d-kube-api-access-tkkxm\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322347 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/857289f7-30a2-42ed-8004-188b9e0a1626-signing-cabundle\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322373 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvjcd\" (UniqueName: \"kubernetes.io/projected/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-kube-api-access-rvjcd\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322397 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ba793aa-5158-46a3-8ea6-1e91389524d1-config-volume\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0debf7-f7f7-4445-b5be-fa96774cbcf1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b6msk\" (UID: \"cc0debf7-f7f7-4445-b5be-fa96774cbcf1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322456 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mqsr\" (UniqueName: \"kubernetes.io/projected/87d27326-3c94-424c-af7e-6defb922382d-kube-api-access-9mqsr\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76a2fafe-8637-4067-a89b-cf6756ba3e09-srv-cert\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322510 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ad1a76d0-a604-4c89-921d-aed33f60fd01-webhook-cert\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322533 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-csi-data-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-plugins-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc5p6\" (UniqueName: \"kubernetes.io/projected/fe723d30-456e-491e-84ed-141775f2592d-kube-api-access-wc5p6\") pod \"ingress-canary-2r4nc\" (UID: \"fe723d30-456e-491e-84ed-141775f2592d\") " pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322624 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-registration-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322662 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpxcz\" (UniqueName: \"kubernetes.io/projected/76a2fafe-8637-4067-a89b-cf6756ba3e09-kube-api-access-lpxcz\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322753 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-socket-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b32adce4-0f8a-433e-8643-a9b35e02fd8d-service-ca-bundle\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322815 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ndx2\" (UniqueName: \"kubernetes.io/projected/ad1a76d0-a604-4c89-921d-aed33f60fd01-kube-api-access-9ndx2\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322839 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ba793aa-5158-46a3-8ea6-1e91389524d1-secret-volume\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.322869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.324096 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9afc51-40c4-46e2-8fdf-2d29f7520271-config\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.324370 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hb8lk"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.325079 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.325331 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-images\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.325419 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-mountpoint-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.326375 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65f8689-77b0-4873-af9d-d82f5ec79c53-config-volume\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.327146 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-664pd\" (UniqueName: \"kubernetes.io/projected/556ada8d-ee27-46a9-94cd-a0ec52b077aa-kube-api-access-664pd\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.327230 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-proxy-tls\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.327493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/87d27326-3c94-424c-af7e-6defb922382d-srv-cert\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.327647 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/857289f7-30a2-42ed-8004-188b9e0a1626-signing-key\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.328300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76a2fafe-8637-4067-a89b-cf6756ba3e09-profile-collector-cert\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.328343 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-auth-proxy-config\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.329394 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.329525 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f9afc51-40c4-46e2-8fdf-2d29f7520271-serving-cert\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.329646 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-registration-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.329774 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kr5ql"] Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.330143 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:46.830124365 +0000 UTC m=+156.184039336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.332873 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17dcfe6a-c319-4ac4-82a9-407601bf589b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9r6ts\" (UID: \"17dcfe6a-c319-4ac4-82a9-407601bf589b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.333154 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0debf7-f7f7-4445-b5be-fa96774cbcf1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b6msk\" (UID: \"cc0debf7-f7f7-4445-b5be-fa96774cbcf1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.333402 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/87d27326-3c94-424c-af7e-6defb922382d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.333541 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-socket-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.333725 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77e064ce-ee89-43e9-b600-a5ecf951cead-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.334298 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77e064ce-ee89-43e9-b600-a5ecf951cead-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.335021 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-default-certificate\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.335075 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.336093 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/dba538a8-b800-4492-b281-b4eb25873b71-certs\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.339900 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-plugins-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.339901 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-csi-data-dir\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.340928 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/857289f7-30a2-42ed-8004-188b9e0a1626-signing-cabundle\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.341878 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppfdh\" (UniqueName: \"kubernetes.io/projected/fd7496df-c90b-454b-b107-6345db7738ca-kube-api-access-ppfdh\") pod \"dns-operator-744455d44c-m7d4b\" (UID: \"fd7496df-c90b-454b-b107-6345db7738ca\") " pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.341884 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-metrics-certs\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.344022 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76a2fafe-8637-4067-a89b-cf6756ba3e09-srv-cert\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.345444 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f65f8689-77b0-4873-af9d-d82f5ec79c53-metrics-tls\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.347362 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b32adce4-0f8a-433e-8643-a9b35e02fd8d-stats-auth\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.362577 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sprf\" (UniqueName: \"kubernetes.io/projected/483f2b9f-6fcd-4569-ab81-a6f265083617-kube-api-access-4sprf\") pod \"openshift-config-operator-7777fb866f-bjkt9\" (UID: \"483f2b9f-6fcd-4569-ab81-a6f265083617\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.368775 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ba793aa-5158-46a3-8ea6-1e91389524d1-config-volume\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.369198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ad1a76d0-a604-4c89-921d-aed33f60fd01-tmpfs\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.369367 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b32adce4-0f8a-433e-8643-a9b35e02fd8d-service-ca-bundle\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.371752 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ad1a76d0-a604-4c89-921d-aed33f60fd01-apiservice-cert\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.371789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ad1a76d0-a604-4c89-921d-aed33f60fd01-webhook-cert\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.371812 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ba793aa-5158-46a3-8ea6-1e91389524d1-secret-volume\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.376623 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fe723d30-456e-491e-84ed-141775f2592d-cert\") pod \"ingress-canary-2r4nc\" (UID: \"fe723d30-456e-491e-84ed-141775f2592d\") " pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.377590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c56509df-b9b8-4857-929f-2a2667dc9e58-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-djlfq\" (UID: \"c56509df-b9b8-4857-929f-2a2667dc9e58\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.377679 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/dba538a8-b800-4492-b281-b4eb25873b71-node-bootstrap-token\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: W1008 15:15:46.381032 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54eb5003_d347_47aa_b14a_391d1bd74bdb.slice/crio-83db08a5a71163ccf4657d137e1f7b32d7ce3950bb5a0cae60a6c415d1f0efe4 WatchSource:0}: Error finding container 83db08a5a71163ccf4657d137e1f7b32d7ce3950bb5a0cae60a6c415d1f0efe4: Status 404 returned error can't find the container with id 83db08a5a71163ccf4657d137e1f7b32d7ce3950bb5a0cae60a6c415d1f0efe4 Oct 08 15:15:46 crc kubenswrapper[4945]: W1008 15:15:46.386214 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79d1c40a_ac76_4f69_94da_b701c1804a54.slice/crio-a28a7c4658cdcd9a5ed0e4467ac4bd351329a64d7446b15728cbddabfb1be192 WatchSource:0}: Error finding container a28a7c4658cdcd9a5ed0e4467ac4bd351329a64d7446b15728cbddabfb1be192: Status 404 returned error can't find the container with id a28a7c4658cdcd9a5ed0e4467ac4bd351329a64d7446b15728cbddabfb1be192 Oct 08 15:15:46 crc kubenswrapper[4945]: W1008 15:15:46.391941 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6df6697f_bec8_4b87_a0ef_7e71c9a0b360.slice/crio-ca92baacaa8eaf12592e80a2ff5dcea43a3e725bc853f2be507c362ea791904c WatchSource:0}: Error finding container ca92baacaa8eaf12592e80a2ff5dcea43a3e725bc853f2be507c362ea791904c: Status 404 returned error can't find the container with id ca92baacaa8eaf12592e80a2ff5dcea43a3e725bc853f2be507c362ea791904c Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.395659 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpgq2\" (UniqueName: \"kubernetes.io/projected/f973b44d-6296-408b-9d14-31d56fdd1021-kube-api-access-hpgq2\") pod \"control-plane-machine-set-operator-78cbb6b69f-mn52j\" (UID: \"f973b44d-6296-408b-9d14-31d56fdd1021\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.411343 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/556ada8d-ee27-46a9-94cd-a0ec52b077aa-bound-sa-token\") pod \"ingress-operator-5b745b69d9-82pkj\" (UID: \"556ada8d-ee27-46a9-94cd-a0ec52b077aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.417409 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.423852 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.424348 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:46.924242617 +0000 UTC m=+156.278157518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.425353 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.425506 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:46.925483968 +0000 UTC m=+156.279399039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.434526 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4ndj\" (UniqueName: \"kubernetes.io/projected/074308ce-e428-4d80-a9f3-e4ff0cd4cc4b-kube-api-access-v4ndj\") pod \"etcd-operator-b45778765-dl6kq\" (UID: \"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.448085 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.452802 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-bound-sa-token\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.469183 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.495621 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw27x\" (UniqueName: \"kubernetes.io/projected/e76deaf4-7012-4889-87bf-1310e42a6f46-kube-api-access-tw27x\") pod \"marketplace-operator-79b997595-m5fwt\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.513787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.515164 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64dlt\" (UniqueName: \"kubernetes.io/projected/f65f8689-77b0-4873-af9d-d82f5ec79c53-kube-api-access-64dlt\") pod \"dns-default-vkzx6\" (UID: \"f65f8689-77b0-4873-af9d-d82f5ec79c53\") " pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.527063 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.527314 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.027287182 +0000 UTC m=+156.381202083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.527663 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.528158 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.028137423 +0000 UTC m=+156.382052324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.534176 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbztx\" (UniqueName: \"kubernetes.io/projected/77e064ce-ee89-43e9-b600-a5ecf951cead-kube-api-access-hbztx\") pod \"kube-storage-version-migrator-operator-b67b599dd-jkw5g\" (UID: \"77e064ce-ee89-43e9-b600-a5ecf951cead\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.536014 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.551732 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.556387 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v59d\" (UniqueName: \"kubernetes.io/projected/17dcfe6a-c319-4ac4-82a9-407601bf589b-kube-api-access-5v59d\") pod \"multus-admission-controller-857f4d67dd-9r6ts\" (UID: \"17dcfe6a-c319-4ac4-82a9-407601bf589b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.560220 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.573278 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.573363 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.575738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6d6x\" (UniqueName: \"kubernetes.io/projected/a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7-kube-api-access-r6d6x\") pod \"csi-hostpathplugin-6trg2\" (UID: \"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.587167 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.591534 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.601582 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.602260 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2z64\" (UniqueName: \"kubernetes.io/projected/86210512-22db-4bf6-876b-f10593c85e02-kube-api-access-b2z64\") pod \"migrator-59844c95c7-k8w22\" (UID: \"86210512-22db-4bf6-876b-f10593c85e02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.611762 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.621157 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.630714 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.631393 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.131374443 +0000 UTC m=+156.485289344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.631826 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.632819 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b5mwn"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.635381 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h6j2\" (UniqueName: \"kubernetes.io/projected/0f9afc51-40c4-46e2-8fdf-2d29f7520271-kube-api-access-6h6j2\") pod \"service-ca-operator-777779d784-62xzf\" (UID: \"0f9afc51-40c4-46e2-8fdf-2d29f7520271\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.638492 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc5p6\" (UniqueName: \"kubernetes.io/projected/fe723d30-456e-491e-84ed-141775f2592d-kube-api-access-wc5p6\") pod \"ingress-canary-2r4nc\" (UID: \"fe723d30-456e-491e-84ed-141775f2592d\") " pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.653831 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpxcz\" (UniqueName: \"kubernetes.io/projected/76a2fafe-8637-4067-a89b-cf6756ba3e09-kube-api-access-lpxcz\") pod \"catalog-operator-68c6474976-tjzkw\" (UID: \"76a2fafe-8637-4067-a89b-cf6756ba3e09\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.658328 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.682958 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ndx2\" (UniqueName: \"kubernetes.io/projected/ad1a76d0-a604-4c89-921d-aed33f60fd01-kube-api-access-9ndx2\") pod \"packageserver-d55dfcdfc-8rb2p\" (UID: \"ad1a76d0-a604-4c89-921d-aed33f60fd01\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.732973 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.733283 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.23327216 +0000 UTC m=+156.587187061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.747751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkkxm\" (UniqueName: \"kubernetes.io/projected/b32adce4-0f8a-433e-8643-a9b35e02fd8d-kube-api-access-tkkxm\") pod \"router-default-5444994796-pr9cp\" (UID: \"b32adce4-0f8a-433e-8643-a9b35e02fd8d\") " pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.748318 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkkxp\" (UniqueName: \"kubernetes.io/projected/857289f7-30a2-42ed-8004-188b9e0a1626-kube-api-access-wkkxp\") pod \"service-ca-9c57cc56f-tnlpm\" (UID: \"857289f7-30a2-42ed-8004-188b9e0a1626\") " pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.750631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mqsr\" (UniqueName: \"kubernetes.io/projected/87d27326-3c94-424c-af7e-6defb922382d-kube-api-access-9mqsr\") pod \"olm-operator-6b444d44fb-fdfbw\" (UID: \"87d27326-3c94-424c-af7e-6defb922382d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.752764 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvjcd\" (UniqueName: \"kubernetes.io/projected/f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64-kube-api-access-rvjcd\") pod \"machine-config-operator-74547568cd-79mrb\" (UID: \"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.792764 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mfmk\" (UniqueName: \"kubernetes.io/projected/cc0debf7-f7f7-4445-b5be-fa96774cbcf1-kube-api-access-7mfmk\") pod \"package-server-manager-789f6589d5-b6msk\" (UID: \"cc0debf7-f7f7-4445-b5be-fa96774cbcf1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.811064 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" event={"ID":"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193","Type":"ContainerStarted","Data":"96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.812200 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.817179 4945 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-477zg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.817235 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" podUID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.819247 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" event={"ID":"8db81fa8-6d99-4529-a1a8-c573752dfb18","Type":"ContainerStarted","Data":"fc3bd2630277d9e53ec5b55b53e11506ad3d22d14933b6aae1b5adab6ac3136d"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.822308 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ft2\" (UniqueName: \"kubernetes.io/projected/2ba793aa-5158-46a3-8ea6-1e91389524d1-kube-api-access-j9ft2\") pod \"collect-profiles-29332275-wplnp\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.838826 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" event={"ID":"536ab60e-f81d-4b67-9528-fe765502f8e0","Type":"ContainerStarted","Data":"0fa8b5e8a2ad9cb1ce40f7b5a1de44da9ec09a79413b80afb2b4aabc893c14d5"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.838975 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.840955 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.340933981 +0000 UTC m=+156.694848892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.858821 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" event={"ID":"79d1c40a-ac76-4f69-94da-b701c1804a54","Type":"ContainerStarted","Data":"a28a7c4658cdcd9a5ed0e4467ac4bd351329a64d7446b15728cbddabfb1be192"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.872732 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.872900 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.873213 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.873499 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.873507 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.873703 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2r4nc" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.883779 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.891697 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.892655 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" event={"ID":"acf1d244-e7ed-41ec-8bc4-ef88aedb1411","Type":"ContainerStarted","Data":"ca80797fd34038b476793f713db5e80f33fa0a2a0f8835996591f1e4d47de82b"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.897166 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxxnw\" (UniqueName: \"kubernetes.io/projected/dba538a8-b800-4492-b281-b4eb25873b71-kube-api-access-bxxnw\") pod \"machine-config-server-7k5pl\" (UID: \"dba538a8-b800-4492-b281-b4eb25873b71\") " pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.926264 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9"] Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.930006 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.931439 4945 generic.go:334] "Generic (PLEG): container finished" podID="8e75c21c-179f-4c78-b1f4-dab40ee201fb" containerID="beccefd88ba7254805407e0d1cea6151cde8a8a98d0cec7f203d5370bd33aee4" exitCode=0 Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.931521 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" event={"ID":"8e75c21c-179f-4c78-b1f4-dab40ee201fb","Type":"ContainerDied","Data":"beccefd88ba7254805407e0d1cea6151cde8a8a98d0cec7f203d5370bd33aee4"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.934006 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" event={"ID":"805c4266-7d9c-4a26-af0d-9d25ca406fb8","Type":"ContainerStarted","Data":"b32d1e7554f75ff4c5251ed2f0385f007cc1bc6d8068bf66a7dffae56fc76bcb"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.935266 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" event={"ID":"cdac1f7c-1f67-43f8-8801-4014a03f9a14","Type":"ContainerStarted","Data":"56083f58b95e79dd1bb8647a1c885f806ecf581c5da0867fe0338d13d728d18f"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.936321 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" event={"ID":"2dca32bb-bf0b-4632-8fe5-af683164904a","Type":"ContainerStarted","Data":"1cbec38240330eb865b641b9bd49fbce9691b30c65c8b83a7fa1ffaa894b861b"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.940215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.940882 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:46 crc kubenswrapper[4945]: E1008 15:15:46.941154 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.441138795 +0000 UTC m=+156.795053766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.941274 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" event={"ID":"b5f113bd-0207-40ce-9793-c7b0c901b885","Type":"ContainerStarted","Data":"53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.941399 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.942985 4945 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xghvq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.943060 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" podUID="b5f113bd-0207-40ce-9793-c7b0c901b885" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.946468 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" event={"ID":"075d448b-cf3c-47e4-a9bb-92d9a8295a2d","Type":"ContainerStarted","Data":"b8b4b07e500eb0340daa7f5f3e7bbc7b99a1b86ce37f9be719a18df11d2dffdb"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.950288 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.968277 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kr5ql" event={"ID":"6df6697f-bec8-4b87-a0ef-7e71c9a0b360","Type":"ContainerStarted","Data":"ca92baacaa8eaf12592e80a2ff5dcea43a3e725bc853f2be507c362ea791904c"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.971052 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b5mwn" event={"ID":"9b6a0a68-fe4e-4f50-af67-6d96384b1064","Type":"ContainerStarted","Data":"e2b28db5b7433bc3bb189f4fb8c2b0237f303769648a182537eb50c5270cdd2d"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.973852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" event={"ID":"44b02303-2b9b-448e-86e5-ed9bc3a59eec","Type":"ContainerStarted","Data":"6aa12e91b863b6639cf64bd3e3babd94902ee4e6264b2c86e1d1b7a616de1a09"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.975604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" event={"ID":"ec002292-f388-497c-b114-fb7818730f4d","Type":"ContainerStarted","Data":"c7dfe2c01aff6214067984390638202b08489d1d58ba0d2946b09ffa97978a4d"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.975642 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" event={"ID":"ec002292-f388-497c-b114-fb7818730f4d","Type":"ContainerStarted","Data":"89e87395e24a5bf4318cd9427a61499c02f1529d97cec81f5c3e2a1675d4462c"} Oct 08 15:15:46 crc kubenswrapper[4945]: I1008 15:15:46.989129 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.009356 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hb8lk" event={"ID":"54eb5003-d347-47aa-b14a-391d1bd74bdb","Type":"ContainerStarted","Data":"83db08a5a71163ccf4657d137e1f7b32d7ce3950bb5a0cae60a6c415d1f0efe4"} Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.041672 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.043068 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.543043151 +0000 UTC m=+156.896958052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.149812 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.150501 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.650473706 +0000 UTC m=+157.004388617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.172635 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7k5pl" Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.256484 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.256602 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.756579848 +0000 UTC m=+157.110494749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.256986 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.257394 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.757384638 +0000 UTC m=+157.111299539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.357503 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.357815 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.857798138 +0000 UTC m=+157.211713039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.389382 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.459026 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.459398 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:47.959383636 +0000 UTC m=+157.313298537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.517950 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.550789 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-m7d4b"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.552754 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.559594 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.559736 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.059706024 +0000 UTC m=+157.413620935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.560570 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.561017 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.061002466 +0000 UTC m=+157.414917367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.575235 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vkzx6"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.586570 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dl6kq"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.593865 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m5fwt"] Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.670464 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.17040418 +0000 UTC m=+157.524319091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.668675 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.674023 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.674598 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.174580284 +0000 UTC m=+157.528495185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.675600 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.694433 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9r6ts"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.710024 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw"] Oct 08 15:15:47 crc kubenswrapper[4945]: W1008 15:15:47.713435 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86210512_22db_4bf6_876b_f10593c85e02.slice/crio-391832c3d204a561c9dcf0767ceec92b051c1edcf9a48da606133b6996b395ae WatchSource:0}: Error finding container 391832c3d204a561c9dcf0767ceec92b051c1edcf9a48da606133b6996b395ae: Status 404 returned error can't find the container with id 391832c3d204a561c9dcf0767ceec92b051c1edcf9a48da606133b6996b395ae Oct 08 15:15:47 crc kubenswrapper[4945]: W1008 15:15:47.716311 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17dcfe6a_c319_4ac4_82a9_407601bf589b.slice/crio-95b9cc681ce305af170f169fc7c92ec85ec636bc941c4acdbf8169d2498585cd WatchSource:0}: Error finding container 95b9cc681ce305af170f169fc7c92ec85ec636bc941c4acdbf8169d2498585cd: Status 404 returned error can't find the container with id 95b9cc681ce305af170f169fc7c92ec85ec636bc941c4acdbf8169d2498585cd Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.775094 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.775262 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.27522665 +0000 UTC m=+157.629141551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.775324 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.775651 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.27563757 +0000 UTC m=+157.629552471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.879747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.880168 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.380147982 +0000 UTC m=+157.734062883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.882490 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-62xzf"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.886431 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.895413 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.901428 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.921733 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tnlpm"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.965746 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2r4nc"] Oct 08 15:15:47 crc kubenswrapper[4945]: W1008 15:15:47.973206 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87d27326_3c94_424c_af7e_6defb922382d.slice/crio-11561ea9d28ee17ed6ecc3b5f6bbf3c7f3ff79aa86f5e2feb3daa669e1507af7 WatchSource:0}: Error finding container 11561ea9d28ee17ed6ecc3b5f6bbf3c7f3ff79aa86f5e2feb3daa669e1507af7: Status 404 returned error can't find the container with id 11561ea9d28ee17ed6ecc3b5f6bbf3c7f3ff79aa86f5e2feb3daa669e1507af7 Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.975352 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6trg2"] Oct 08 15:15:47 crc kubenswrapper[4945]: I1008 15:15:47.980827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:47 crc kubenswrapper[4945]: E1008 15:15:47.981100 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.481087934 +0000 UTC m=+157.835002835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: W1008 15:15:48.023215 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe723d30_456e_491e_84ed_141775f2592d.slice/crio-5be08e0e7020569c03ef964d5e95395108e9439a2c9e8c126872b7eb163b3ffe WatchSource:0}: Error finding container 5be08e0e7020569c03ef964d5e95395108e9439a2c9e8c126872b7eb163b3ffe: Status 404 returned error can't find the container with id 5be08e0e7020569c03ef964d5e95395108e9439a2c9e8c126872b7eb163b3ffe Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.036217 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" event={"ID":"c56509df-b9b8-4857-929f-2a2667dc9e58","Type":"ContainerStarted","Data":"b5a935be164ff5fed5fcc186336347ccd3afb47d75d464c5b74938ad78ad0e51"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.036253 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" event={"ID":"17dcfe6a-c319-4ac4-82a9-407601bf589b","Type":"ContainerStarted","Data":"95b9cc681ce305af170f169fc7c92ec85ec636bc941c4acdbf8169d2498585cd"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.040340 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" event={"ID":"556ada8d-ee27-46a9-94cd-a0ec52b077aa","Type":"ContainerStarted","Data":"d6876a2ccd904e921c401d01a68cb935f4325f001a4c1df6913b6908b17055cb"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.044425 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb"] Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.058263 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" event={"ID":"ad1a76d0-a604-4c89-921d-aed33f60fd01","Type":"ContainerStarted","Data":"d5b5a41302b7fe3b987480122a278c9557a10e11e3855b188dddabb92e2e4a63"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.060481 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp"] Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.060600 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" event={"ID":"79d1c40a-ac76-4f69-94da-b701c1804a54","Type":"ContainerStarted","Data":"ee70686c95215cf975537f4176e9c953e4d984dc2ee930019f3b209eb08a4116"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.061899 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" event={"ID":"76a2fafe-8637-4067-a89b-cf6756ba3e09","Type":"ContainerStarted","Data":"04c687a42618ee9712bb5d9a7a591ae43acabc6fa80bef28e91ad2fac35eee8a"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.066747 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" event={"ID":"0f9afc51-40c4-46e2-8fdf-2d29f7520271","Type":"ContainerStarted","Data":"3f5b93cd27372fd1f6de9d9f3c170b57f480ec5705b13b119b1a200219980221"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.067728 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" event={"ID":"86210512-22db-4bf6-876b-f10593c85e02","Type":"ContainerStarted","Data":"391832c3d204a561c9dcf0767ceec92b051c1edcf9a48da606133b6996b395ae"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.068914 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" event={"ID":"cc0debf7-f7f7-4445-b5be-fa96774cbcf1","Type":"ContainerStarted","Data":"733ed6befcc957e42bbda41f381186cbb54a3295164fd3a69bc24b16d41d78b8"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.070080 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" event={"ID":"8ef73f3d-3f61-499c-8806-cf041851986c","Type":"ContainerStarted","Data":"6462141b905214040df00b9f2c1d49341ec8923b4b6f7b81b4cba118e102da66"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.071404 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" event={"ID":"cdac1f7c-1f67-43f8-8801-4014a03f9a14","Type":"ContainerStarted","Data":"4f0486dde90864b27592d5a6ab5afe992fd21faa8a2c37684773e98d118c9b1f"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.072268 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" event={"ID":"fd7496df-c90b-454b-b107-6345db7738ca","Type":"ContainerStarted","Data":"9cfeaacfbd0e45b1a0956ea9d778e23567eda0378476c63827cc09422e18bc69"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.073061 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" event={"ID":"e76deaf4-7012-4889-87bf-1310e42a6f46","Type":"ContainerStarted","Data":"cac52278a6d6e278a45b93bc8f09be36005d6d227bd18251bd86b6fac185a8b5"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.073806 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" event={"ID":"77e064ce-ee89-43e9-b600-a5ecf951cead","Type":"ContainerStarted","Data":"cc633df91e397368aa15450a2eb0d320d6e26936ee5bf2a87ab0f2da8df2b675"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.076341 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b5mwn" event={"ID":"9b6a0a68-fe4e-4f50-af67-6d96384b1064","Type":"ContainerStarted","Data":"c08f4623c7554f5f4bd20c2f58fae5697f3ee8f2d4650bed9b4263e2c87cc967"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.081627 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.081874 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.581847483 +0000 UTC m=+157.935762394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.082123 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.082430 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.582414436 +0000 UTC m=+157.936329417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.082876 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" event={"ID":"536ab60e-f81d-4b67-9528-fe765502f8e0","Type":"ContainerStarted","Data":"599367a547d98148d3194ee0438d70d6b7d8dafcc331fd37c2b5f8f2dda96a53"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.083755 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" event={"ID":"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b","Type":"ContainerStarted","Data":"c5f643c137e520fbdafdfd2229bcbc38b420163d26ec0b3e53215ae8f36be890"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.086999 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" event={"ID":"2dca32bb-bf0b-4632-8fe5-af683164904a","Type":"ContainerStarted","Data":"69e7610fcd3a0d25a4f08527f10a5246214ab303a6d46ce44ad0fd0009ba9cbe"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.088423 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7k5pl" event={"ID":"dba538a8-b800-4492-b281-b4eb25873b71","Type":"ContainerStarted","Data":"58e37a0555d211009ce0894aa34bd4dfcf647ab08859be49a0d492a210887380"} Oct 08 15:15:48 crc kubenswrapper[4945]: W1008 15:15:48.091244 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf78b6bd7_d6db_4d9f_b7a0_3e970cda0d64.slice/crio-457bb65ce22a4871cecbad286315b225796df3d06c29c091fd91eeff75ec1fa1 WatchSource:0}: Error finding container 457bb65ce22a4871cecbad286315b225796df3d06c29c091fd91eeff75ec1fa1: Status 404 returned error can't find the container with id 457bb65ce22a4871cecbad286315b225796df3d06c29c091fd91eeff75ec1fa1 Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.091272 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" event={"ID":"8db81fa8-6d99-4529-a1a8-c573752dfb18","Type":"ContainerStarted","Data":"05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e"} Oct 08 15:15:48 crc kubenswrapper[4945]: W1008 15:15:48.091632 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ba793aa_5158_46a3_8ea6_1e91389524d1.slice/crio-25990d6267527ec1414b1c6a15fb1738d1e4cb2efa57c0a1f973e1617fc5e56f WatchSource:0}: Error finding container 25990d6267527ec1414b1c6a15fb1738d1e4cb2efa57c0a1f973e1617fc5e56f: Status 404 returned error can't find the container with id 25990d6267527ec1414b1c6a15fb1738d1e4cb2efa57c0a1f973e1617fc5e56f Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.092833 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pr9cp" event={"ID":"b32adce4-0f8a-433e-8643-a9b35e02fd8d","Type":"ContainerStarted","Data":"e6527328f7d694ee719d56d2a219de2e0f06498e2a4ebe6331344a76503783f2"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.095041 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" event={"ID":"483f2b9f-6fcd-4569-ab81-a6f265083617","Type":"ContainerStarted","Data":"abc2c14a7036894e5408cff99ba29e00a11112c2f78e4a0116fd82c3832a28dc"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.096149 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hb8lk" event={"ID":"54eb5003-d347-47aa-b14a-391d1bd74bdb","Type":"ContainerStarted","Data":"d4a3168436dfbe3143ff00dda0bc8d66f8f846030d827a0ac108c0318e480069"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.097367 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" event={"ID":"951bacb4-e545-4279-ad11-b791aed0aa1e","Type":"ContainerStarted","Data":"a947256334237daad9bfc639915036520bbb5223dca5587d18745ac73dfe6d3d"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.104559 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" event={"ID":"805c4266-7d9c-4a26-af0d-9d25ca406fb8","Type":"ContainerStarted","Data":"00596d10265e82f1d857f3294fe249f9d5c8ec5d66267c0d666dbe7b53c68146"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.111799 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vkzx6" event={"ID":"f65f8689-77b0-4873-af9d-d82f5ec79c53","Type":"ContainerStarted","Data":"e9679c78aefcf12b2d63d2a41169b9efc454cba220af420ecccedf07a3b4ae5d"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.113143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" event={"ID":"87d27326-3c94-424c-af7e-6defb922382d","Type":"ContainerStarted","Data":"11561ea9d28ee17ed6ecc3b5f6bbf3c7f3ff79aa86f5e2feb3daa669e1507af7"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.114028 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" event={"ID":"f973b44d-6296-408b-9d14-31d56fdd1021","Type":"ContainerStarted","Data":"500a36283f4291a32128a7d82d31d713fecbed03ea5875016785e0958e9b6cf5"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.115725 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kr5ql" event={"ID":"6df6697f-bec8-4b87-a0ef-7e71c9a0b360","Type":"ContainerStarted","Data":"ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.116856 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" event={"ID":"857289f7-30a2-42ed-8004-188b9e0a1626","Type":"ContainerStarted","Data":"108f5ca0d021a32b1e7ad4054231257180c8592ad38967b391d3e748713f0f77"} Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.117918 4945 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xghvq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.117951 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" podUID="b5f113bd-0207-40ce-9793-c7b0c901b885" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.118939 4945 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-477zg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.118967 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" podUID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.183496 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.187068 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.687023371 +0000 UTC m=+158.040938272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.285739 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.286021 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.786009895 +0000 UTC m=+158.139924796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.324753 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j72zk" podStartSLOduration=128.324733172 podStartE2EDuration="2m8.324733172s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:48.272845105 +0000 UTC m=+157.626760006" watchObservedRunningTime="2025-10-08 15:15:48.324733172 +0000 UTC m=+157.678648103" Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.326083 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5k2tz" podStartSLOduration=129.326075706 podStartE2EDuration="2m9.326075706s" podCreationTimestamp="2025-10-08 15:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:48.318324983 +0000 UTC m=+157.672239894" watchObservedRunningTime="2025-10-08 15:15:48.326075706 +0000 UTC m=+157.679990597" Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.362899 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" podStartSLOduration=128.362881306 podStartE2EDuration="2m8.362881306s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:48.361234855 +0000 UTC m=+157.715149756" watchObservedRunningTime="2025-10-08 15:15:48.362881306 +0000 UTC m=+157.716796207" Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.391340 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" podStartSLOduration=128.391324947 podStartE2EDuration="2m8.391324947s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:48.389611864 +0000 UTC m=+157.743526765" watchObservedRunningTime="2025-10-08 15:15:48.391324947 +0000 UTC m=+157.745239848" Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.391652 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.391852 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.891836059 +0000 UTC m=+158.245750960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.391913 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.392251 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.89224472 +0000 UTC m=+158.246159621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.493175 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.494364 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:48.994346382 +0000 UTC m=+158.348261283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.595702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.596371 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.09635958 +0000 UTC m=+158.450274471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.700699 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.700862 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.200838102 +0000 UTC m=+158.554753003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.701015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.701313 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.201301543 +0000 UTC m=+158.555216444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.802356 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.802723 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.302676887 +0000 UTC m=+158.656591828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.802925 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.803503 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.303487097 +0000 UTC m=+158.657401998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.904511 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.904678 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.404657156 +0000 UTC m=+158.758572057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:48 crc kubenswrapper[4945]: I1008 15:15:48.904760 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:48 crc kubenswrapper[4945]: E1008 15:15:48.905118 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.405092637 +0000 UTC m=+158.759007538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.013704 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.013850 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.513832824 +0000 UTC m=+158.867747725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.013949 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.014258 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.514250185 +0000 UTC m=+158.868165086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.115241 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.115570 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.615555066 +0000 UTC m=+158.969469967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.121195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" event={"ID":"2ba793aa-5158-46a3-8ea6-1e91389524d1","Type":"ContainerStarted","Data":"25990d6267527ec1414b1c6a15fb1738d1e4cb2efa57c0a1f973e1617fc5e56f"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.121992 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" event={"ID":"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7","Type":"ContainerStarted","Data":"0c3fae56882315a3c897b09d8fd5b3f4639c030a52f765422f267bbb81f878a7"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.123893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7k5pl" event={"ID":"dba538a8-b800-4492-b281-b4eb25873b71","Type":"ContainerStarted","Data":"f5c15af685bc1011261fb170fabe4ef8d79890f5279013a370f2ae31f70c73a1"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.124779 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2r4nc" event={"ID":"fe723d30-456e-491e-84ed-141775f2592d","Type":"ContainerStarted","Data":"5be08e0e7020569c03ef964d5e95395108e9439a2c9e8c126872b7eb163b3ffe"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.126965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" event={"ID":"c56509df-b9b8-4857-929f-2a2667dc9e58","Type":"ContainerStarted","Data":"76019704f0a2cc12bf4ae30d389aae6e646ec2e56edbfbce497654614b0f5c1d"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.128599 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" event={"ID":"483f2b9f-6fcd-4569-ab81-a6f265083617","Type":"ContainerStarted","Data":"475b460f936c16f8a59c258633131b456c31e0f0d39d484c08e20b544797fe23"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.132343 4945 generic.go:334] "Generic (PLEG): container finished" podID="44b02303-2b9b-448e-86e5-ed9bc3a59eec" containerID="50ae9ec7074c2bec7618194144d7120544ea936bc76b6203070c23d3dad6a9e5" exitCode=0 Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.132401 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" event={"ID":"44b02303-2b9b-448e-86e5-ed9bc3a59eec","Type":"ContainerDied","Data":"50ae9ec7074c2bec7618194144d7120544ea936bc76b6203070c23d3dad6a9e5"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.133463 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" event={"ID":"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64","Type":"ContainerStarted","Data":"457bb65ce22a4871cecbad286315b225796df3d06c29c091fd91eeff75ec1fa1"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.134557 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pr9cp" event={"ID":"b32adce4-0f8a-433e-8643-a9b35e02fd8d","Type":"ContainerStarted","Data":"37e5be8673599f7d8047036262c5fb4b56021ee758def3af04210dcd546610d1"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.135993 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" event={"ID":"951bacb4-e545-4279-ad11-b791aed0aa1e","Type":"ContainerStarted","Data":"78a9515ed2a5ab47f64e5aeff5bfd09084a48d1aa09e544ffb86a48312d9fca9"} Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.136438 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.138793 4945 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-d8prt container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.138849 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" podUID="8db81fa8-6d99-4529-a1a8-c573752dfb18" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.145021 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.153820 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" podStartSLOduration=129.153798562 podStartE2EDuration="2m9.153798562s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:49.153384611 +0000 UTC m=+158.507299522" watchObservedRunningTime="2025-10-08 15:15:49.153798562 +0000 UTC m=+158.507713473" Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.157868 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-95k2h" podStartSLOduration=130.157846254 podStartE2EDuration="2m10.157846254s" podCreationTimestamp="2025-10-08 15:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:48.42743937 +0000 UTC m=+157.781354281" watchObservedRunningTime="2025-10-08 15:15:49.157846254 +0000 UTC m=+158.511761165" Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.176211 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xlgt2" podStartSLOduration=129.176194842 podStartE2EDuration="2m9.176194842s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:49.174042098 +0000 UTC m=+158.527957019" watchObservedRunningTime="2025-10-08 15:15:49.176194842 +0000 UTC m=+158.530109763" Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.184840 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.184910 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.216989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.217917 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.717902294 +0000 UTC m=+159.071817285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.318064 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.318244 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.818218411 +0000 UTC m=+159.172133312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.318444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.318808 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.818799565 +0000 UTC m=+159.172714466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.419404 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.419673 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.919622615 +0000 UTC m=+159.273537566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.420435 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.420966 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:49.920947988 +0000 UTC m=+159.274862929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.522177 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.522381 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.022342033 +0000 UTC m=+159.376256974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.522670 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.522997 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.022981758 +0000 UTC m=+159.376896739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.624380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.624503 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.124484295 +0000 UTC m=+159.478399206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.624931 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.625308 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.125297055 +0000 UTC m=+159.479211966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.725749 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.725901 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.225883039 +0000 UTC m=+159.579797940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.726044 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.726356 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.226348001 +0000 UTC m=+159.580262912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.828776 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.829191 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.329173921 +0000 UTC m=+159.683088822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:49 crc kubenswrapper[4945]: I1008 15:15:49.930003 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:49 crc kubenswrapper[4945]: E1008 15:15:49.930359 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.430347309 +0000 UTC m=+159.784262210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.032643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.032782 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.532762388 +0000 UTC m=+159.886677289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.032956 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.033251 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.533244501 +0000 UTC m=+159.887159402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.134081 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.134238 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.634212154 +0000 UTC m=+159.988127055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.134660 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.134991 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.634979753 +0000 UTC m=+159.988894714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.144688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" event={"ID":"86210512-22db-4bf6-876b-f10593c85e02","Type":"ContainerStarted","Data":"ae156961b9d6460e27c63aec5a5e116e84a5600bfd13e544c87ffb9af2c7cb1e"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.145785 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" event={"ID":"f973b44d-6296-408b-9d14-31d56fdd1021","Type":"ContainerStarted","Data":"e40a36c19aa8ea67e8dedd69846026a25458405a13e505f60c37bfcbb37aece5"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.147098 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" event={"ID":"805c4266-7d9c-4a26-af0d-9d25ca406fb8","Type":"ContainerStarted","Data":"de3605da5ffab5d2e754b63359eaf82f033c172f0aec5278c7809c1bf407c3d2"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.148161 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" event={"ID":"76a2fafe-8637-4067-a89b-cf6756ba3e09","Type":"ContainerStarted","Data":"4f47c09b50836097adc726b99bb4ec720ae8f17b0428b1c9cdcb6ff02ef00736"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.149243 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" event={"ID":"8ef73f3d-3f61-499c-8806-cf041851986c","Type":"ContainerStarted","Data":"9312a2252693ce133b23eaff08ff516d1e7aba5fc097d41e30205463b4939cf5"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.150317 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" event={"ID":"074308ce-e428-4d80-a9f3-e4ff0cd4cc4b","Type":"ContainerStarted","Data":"41d260d7caa0b7d268279d91a457a1dd2128a2737f192f388b1c184ac428fc7a"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.151398 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" event={"ID":"e76deaf4-7012-4889-87bf-1310e42a6f46","Type":"ContainerStarted","Data":"478229d09b8b505553a9ffdf2715f3815d1a0ec868c769a7f07f319ef23f16f0"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.152482 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" event={"ID":"77e064ce-ee89-43e9-b600-a5ecf951cead","Type":"ContainerStarted","Data":"b852492daff21d0f6ad9828a755d0f9139fde52c4e24dd66ff7642e41432fc5c"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.153504 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vkzx6" event={"ID":"f65f8689-77b0-4873-af9d-d82f5ec79c53","Type":"ContainerStarted","Data":"95104aa6387eff97606e506a0c848887f122eb8c3067e1dd66f2760440d0440e"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.154492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" event={"ID":"0f9afc51-40c4-46e2-8fdf-2d29f7520271","Type":"ContainerStarted","Data":"6cde19107d5c2e1ca273264b41aa61f760fb2f0944716afd42f92e7985c644e3"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.155544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" event={"ID":"2ba793aa-5158-46a3-8ea6-1e91389524d1","Type":"ContainerStarted","Data":"2e028ceec119da365ac3a511d19e8c0236278e93a94d7c9f0c81610de0295962"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.156698 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" event={"ID":"556ada8d-ee27-46a9-94cd-a0ec52b077aa","Type":"ContainerStarted","Data":"10a966b5fde0d17c25d2ea905a3a5e74c16fbfdaf5f87d5e0902e52e539fb0a2"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.157755 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" event={"ID":"fd7496df-c90b-454b-b107-6345db7738ca","Type":"ContainerStarted","Data":"50f56716a7a0d7a0eeb145f004a520b3b8a8189a3121368bc9c3df9e5bc12ec2"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.158792 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2r4nc" event={"ID":"fe723d30-456e-491e-84ed-141775f2592d","Type":"ContainerStarted","Data":"8ded44052990608b9dc3e0a4c13c489820578e0f9bf7dfdbab9476c4f3fa669e"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.160390 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" event={"ID":"8e75c21c-179f-4c78-b1f4-dab40ee201fb","Type":"ContainerStarted","Data":"16b301c7571bf579bbc263eda9fa060a5b02dc8461ff7537bc988e01a4e8e149"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.161506 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" event={"ID":"17dcfe6a-c319-4ac4-82a9-407601bf589b","Type":"ContainerStarted","Data":"b0239a9ae67b3578ad6d93c2b11104eb774bb3e916e39c846f57c902a2b70aa6"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.162515 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" event={"ID":"cc0debf7-f7f7-4445-b5be-fa96774cbcf1","Type":"ContainerStarted","Data":"e7660b81a45f311b376f47a2284dfd30c92a8c9f1e4a82ceb64252204466c34b"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.164954 4945 generic.go:334] "Generic (PLEG): container finished" podID="483f2b9f-6fcd-4569-ab81-a6f265083617" containerID="475b460f936c16f8a59c258633131b456c31e0f0d39d484c08e20b544797fe23" exitCode=0 Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.164984 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" event={"ID":"483f2b9f-6fcd-4569-ab81-a6f265083617","Type":"ContainerDied","Data":"475b460f936c16f8a59c258633131b456c31e0f0d39d484c08e20b544797fe23"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.166724 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" event={"ID":"87d27326-3c94-424c-af7e-6defb922382d","Type":"ContainerStarted","Data":"cc1170aa363fdf0e3a784eb9bf516b8e190892140d28ea26e6ef5b653db6a2c9"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.168931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" event={"ID":"ad1a76d0-a604-4c89-921d-aed33f60fd01","Type":"ContainerStarted","Data":"7151a0a31089d9b9048631e099145134dd53661df211a55e2c59cb4ec41de8e7"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.170442 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" event={"ID":"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64","Type":"ContainerStarted","Data":"c96a31d286f1e554802f73602c879b684bd00bb514e82cbbb340500aea9111fe"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.171800 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" event={"ID":"857289f7-30a2-42ed-8004-188b9e0a1626","Type":"ContainerStarted","Data":"1fabcc4852cf19a83143669edf7084f7cbe6b76f7304af79a52c52a83a8d8b17"} Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.203824 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nfdsl" podStartSLOduration=130.203807464 podStartE2EDuration="2m10.203807464s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.202384898 +0000 UTC m=+159.556299819" watchObservedRunningTime="2025-10-08 15:15:50.203807464 +0000 UTC m=+159.557722365" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.236652 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.236762 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.736744196 +0000 UTC m=+160.090659097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.241848 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.245347 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mdsqd" podStartSLOduration=131.245330371 podStartE2EDuration="2m11.245330371s" podCreationTimestamp="2025-10-08 15:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.244584863 +0000 UTC m=+159.598499774" watchObservedRunningTime="2025-10-08 15:15:50.245330371 +0000 UTC m=+159.599245272" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.251640 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.751625429 +0000 UTC m=+160.105540330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.273907 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-hb8lk" podStartSLOduration=130.273884735 podStartE2EDuration="2m10.273884735s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.271506975 +0000 UTC m=+159.625421876" watchObservedRunningTime="2025-10-08 15:15:50.273884735 +0000 UTC m=+159.627799636" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.292697 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djlfq" podStartSLOduration=130.292674844 podStartE2EDuration="2m10.292674844s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.28928667 +0000 UTC m=+159.643201571" watchObservedRunningTime="2025-10-08 15:15:50.292674844 +0000 UTC m=+159.646589745" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.307639 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-b5mwn" podStartSLOduration=130.307622288 podStartE2EDuration="2m10.307622288s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.306465379 +0000 UTC m=+159.660380280" watchObservedRunningTime="2025-10-08 15:15:50.307622288 +0000 UTC m=+159.661537189" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.329604 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-pr9cp" podStartSLOduration=130.329584666 podStartE2EDuration="2m10.329584666s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.324837428 +0000 UTC m=+159.678752329" watchObservedRunningTime="2025-10-08 15:15:50.329584666 +0000 UTC m=+159.683499567" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.346871 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.347255 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.847226497 +0000 UTC m=+160.201141428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.363689 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-429v8" podStartSLOduration=130.363670779 podStartE2EDuration="2m10.363670779s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.346346235 +0000 UTC m=+159.700261136" watchObservedRunningTime="2025-10-08 15:15:50.363670779 +0000 UTC m=+159.717585680" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.365435 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zq9l5" podStartSLOduration=130.365427182 podStartE2EDuration="2m10.365427182s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.36175015 +0000 UTC m=+159.715665051" watchObservedRunningTime="2025-10-08 15:15:50.365427182 +0000 UTC m=+159.719342083" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.389318 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7k5pl" podStartSLOduration=7.389299789 podStartE2EDuration="7.389299789s" podCreationTimestamp="2025-10-08 15:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.389010062 +0000 UTC m=+159.742924963" watchObservedRunningTime="2025-10-08 15:15:50.389299789 +0000 UTC m=+159.743214690" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.406067 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kr5ql" podStartSLOduration=130.406050838 podStartE2EDuration="2m10.406050838s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:50.402725345 +0000 UTC m=+159.756640246" watchObservedRunningTime="2025-10-08 15:15:50.406050838 +0000 UTC m=+159.759965729" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.447967 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.448296 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:50.948285293 +0000 UTC m=+160.302200194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.549021 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.549180 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.049149824 +0000 UTC m=+160.403064725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.549619 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.549909 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.049897833 +0000 UTC m=+160.403812734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.650771 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.650893 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.150871886 +0000 UTC m=+160.504786787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.651198 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.651501 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.151491811 +0000 UTC m=+160.505406712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.752593 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.753045 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.253008409 +0000 UTC m=+160.606923320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.762015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.763353 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.263338386 +0000 UTC m=+160.617253287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.865151 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.865360 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.365331716 +0000 UTC m=+160.719246617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.884625 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.887191 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.887249 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 08 15:15:50 crc kubenswrapper[4945]: I1008 15:15:50.967379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:50 crc kubenswrapper[4945]: E1008 15:15:50.967740 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.467728834 +0000 UTC m=+160.821643735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.068201 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.068367 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.568342299 +0000 UTC m=+160.922257200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.068454 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.068789 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.56878204 +0000 UTC m=+160.922696941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.170049 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.170208 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.670176014 +0000 UTC m=+161.024090925 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.170306 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.170619 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.670607595 +0000 UTC m=+161.024522496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.172796 4945 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-d8prt container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.173076 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" podUID="8db81fa8-6d99-4529-a1a8-c573752dfb18" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.224298 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" podStartSLOduration=51.224266325 podStartE2EDuration="51.224266325s" podCreationTimestamp="2025-10-08 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:51.223399774 +0000 UTC m=+160.577314675" watchObservedRunningTime="2025-10-08 15:15:51.224266325 +0000 UTC m=+160.578181226" Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.249634 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mn52j" podStartSLOduration=131.249616159 podStartE2EDuration="2m11.249616159s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:51.246709916 +0000 UTC m=+160.600624827" watchObservedRunningTime="2025-10-08 15:15:51.249616159 +0000 UTC m=+160.603531070" Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.271467 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.273244 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.773226029 +0000 UTC m=+161.127140930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.375798 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.376459 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.876443219 +0000 UTC m=+161.230358120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.477564 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.477813 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.977765101 +0000 UTC m=+161.331680002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.478346 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:51.978334815 +0000 UTC m=+161.332249716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.478041 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.579638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.579846 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.079825041 +0000 UTC m=+161.433739952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.580003 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.580338 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.080328915 +0000 UTC m=+161.434243816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.680745 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.681161 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.181142614 +0000 UTC m=+161.535057515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.787943 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.788377 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.288359083 +0000 UTC m=+161.642274034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.886283 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.886622 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.888827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.889021 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.388991728 +0000 UTC m=+161.742906639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.889125 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.889481 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.38946992 +0000 UTC m=+161.743384891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:51 crc kubenswrapper[4945]: I1008 15:15:51.990852 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:51 crc kubenswrapper[4945]: E1008 15:15:51.991310 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.491290475 +0000 UTC m=+161.845205386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.092184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.092546 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.592529464 +0000 UTC m=+161.946444435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.193738 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.194301 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.694281138 +0000 UTC m=+162.048196099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.211130 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" event={"ID":"cc0debf7-f7f7-4445-b5be-fa96774cbcf1","Type":"ContainerStarted","Data":"7e0c9b3bfa7b7ee0088cd4b05e48c2234d31934384542d4e57637d13c7c4746b"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.211899 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.215262 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" event={"ID":"483f2b9f-6fcd-4569-ab81-a6f265083617","Type":"ContainerStarted","Data":"284e1b0c9753397f31ff453139ebea2258de764765df6b2425420da5fd474a98"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.215911 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.233885 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" event={"ID":"556ada8d-ee27-46a9-94cd-a0ec52b077aa","Type":"ContainerStarted","Data":"eaaa273240914e3b354364e748b92d6ac90fdba0e558acbcbc16a7d1e3a322cd"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.243549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" event={"ID":"fd7496df-c90b-454b-b107-6345db7738ca","Type":"ContainerStarted","Data":"191c33813598ac4c4ba765b24992dc06a5d7f8375b1908b0c752fb189678460f"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.247151 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" event={"ID":"f78b6bd7-d6db-4d9f-b7a0-3e970cda0d64","Type":"ContainerStarted","Data":"0041ba4d63b83222dc60671909bdc8c5cd565df43d0c4978b6dc4df4c74ec875"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.254523 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" event={"ID":"86210512-22db-4bf6-876b-f10593c85e02","Type":"ContainerStarted","Data":"a2c6ffec6030646dd422697376533fcb169c621aff495410c1af61fa571e432a"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.256931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" event={"ID":"8ef73f3d-3f61-499c-8806-cf041851986c","Type":"ContainerStarted","Data":"a556ae20025f6d0407e7d01b7eca8005482a7f915cccf88ffb2025e13a567709"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.283329 4945 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fdfbw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.283369 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" podUID="87d27326-3c94-424c-af7e-6defb922382d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.286093 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" event={"ID":"44b02303-2b9b-448e-86e5-ed9bc3a59eec","Type":"ContainerStarted","Data":"4860f49e15646341a95cf7aaf2476194469246149ea14b6f2309015baa9ae4db"} Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.286148 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.286160 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.286171 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.286180 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.286632 4945 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-m5fwt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.286658 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" podUID="e76deaf4-7012-4889-87bf-1310e42a6f46" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.295045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.295390 4945 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-tjzkw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.295436 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" podUID="76a2fafe-8637-4067-a89b-cf6756ba3e09" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.295504 4945 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rb2p container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.295522 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" podUID="ad1a76d0-a604-4c89-921d-aed33f60fd01" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.295789 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.795776935 +0000 UTC m=+162.149691826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.340268 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" podStartSLOduration=132.340242765 podStartE2EDuration="2m12.340242765s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.336009269 +0000 UTC m=+161.689924170" watchObservedRunningTime="2025-10-08 15:15:52.340242765 +0000 UTC m=+161.694157776" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.371475 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2r4nc" podStartSLOduration=9.371460086 podStartE2EDuration="9.371460086s" podCreationTimestamp="2025-10-08 15:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.370948973 +0000 UTC m=+161.724863894" watchObservedRunningTime="2025-10-08 15:15:52.371460086 +0000 UTC m=+161.725374987" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.398746 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.398905 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.898879921 +0000 UTC m=+162.252794822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.398992 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.399637 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:52.89963015 +0000 UTC m=+162.253545051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.424370 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjs2d" podStartSLOduration=132.424344617 podStartE2EDuration="2m12.424344617s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.423509057 +0000 UTC m=+161.777423958" watchObservedRunningTime="2025-10-08 15:15:52.424344617 +0000 UTC m=+161.778259518" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.440736 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.441342 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.442963 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.443422 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.455456 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.472641 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" podStartSLOduration=132.472622554 podStartE2EDuration="2m12.472622554s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.472480001 +0000 UTC m=+161.826394902" watchObservedRunningTime="2025-10-08 15:15:52.472622554 +0000 UTC m=+161.826537455" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.500508 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.500748 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.500819 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.500970 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.000956832 +0000 UTC m=+162.354871733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.516826 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c2cpk" podStartSLOduration=132.516810989 podStartE2EDuration="2m12.516810989s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.493915696 +0000 UTC m=+161.847830607" watchObservedRunningTime="2025-10-08 15:15:52.516810989 +0000 UTC m=+161.870725890" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.539927 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dl6kq" podStartSLOduration=132.539906935 podStartE2EDuration="2m12.539906935s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.521460625 +0000 UTC m=+161.875375526" watchObservedRunningTime="2025-10-08 15:15:52.539906935 +0000 UTC m=+161.893821836" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.540856 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jkw5g" podStartSLOduration=132.540850869 podStartE2EDuration="2m12.540850869s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.539570217 +0000 UTC m=+161.893485118" watchObservedRunningTime="2025-10-08 15:15:52.540850869 +0000 UTC m=+161.894765770" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.565061 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-62xzf" podStartSLOduration=132.565047854 podStartE2EDuration="2m12.565047854s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.562525871 +0000 UTC m=+161.916440772" watchObservedRunningTime="2025-10-08 15:15:52.565047854 +0000 UTC m=+161.918962755" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.602453 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.602515 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.602583 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.603172 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.103160206 +0000 UTC m=+162.457075107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.603374 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.675067 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-82pkj" podStartSLOduration=132.675052403 podStartE2EDuration="2m12.675052403s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.597439653 +0000 UTC m=+161.951354554" watchObservedRunningTime="2025-10-08 15:15:52.675052403 +0000 UTC m=+162.028967304" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.675865 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.675925 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-tnlpm" podStartSLOduration=132.675917014 podStartE2EDuration="2m12.675917014s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.674615692 +0000 UTC m=+162.028530593" watchObservedRunningTime="2025-10-08 15:15:52.675917014 +0000 UTC m=+162.029831915" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.705005 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.705361 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.20534573 +0000 UTC m=+162.559260621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.730430 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" podStartSLOduration=132.730407096 podStartE2EDuration="2m12.730407096s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.704484049 +0000 UTC m=+162.058398950" watchObservedRunningTime="2025-10-08 15:15:52.730407096 +0000 UTC m=+162.084321997" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.732443 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" podStartSLOduration=132.732437567 podStartE2EDuration="2m12.732437567s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.729505593 +0000 UTC m=+162.083420504" watchObservedRunningTime="2025-10-08 15:15:52.732437567 +0000 UTC m=+162.086352468" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.767587 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.805993 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.806335 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.306305583 +0000 UTC m=+162.660220484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.898193 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" podStartSLOduration=132.898174979 podStartE2EDuration="2m12.898174979s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.805291868 +0000 UTC m=+162.159206769" watchObservedRunningTime="2025-10-08 15:15:52.898174979 +0000 UTC m=+162.252089880" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.898675 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" podStartSLOduration=132.898670501 podStartE2EDuration="2m12.898670501s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.897362918 +0000 UTC m=+162.251277819" watchObservedRunningTime="2025-10-08 15:15:52.898670501 +0000 UTC m=+162.252585402" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.900433 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:52 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:52 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:52 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.900466 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.907333 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:52 crc kubenswrapper[4945]: E1008 15:15:52.907695 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.407681287 +0000 UTC m=+162.761596188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:52 crc kubenswrapper[4945]: I1008 15:15:52.967412 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" podStartSLOduration=132.967394469 podStartE2EDuration="2m12.967394469s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:52.948923787 +0000 UTC m=+162.302838698" watchObservedRunningTime="2025-10-08 15:15:52.967394469 +0000 UTC m=+162.321309370" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.008630 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.009046 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.509026799 +0000 UTC m=+162.862941780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.109517 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.109660 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.609641844 +0000 UTC m=+162.963556745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.109825 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.110090 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.610080635 +0000 UTC m=+162.963995526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.212585 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.212880 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.712866214 +0000 UTC m=+163.066781115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.303197 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" event={"ID":"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7","Type":"ContainerStarted","Data":"694dba8b3893084ef5df2f1a0740d2491332963f364f2ee495d4019967384816"} Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.311798 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" event={"ID":"8e75c21c-179f-4c78-b1f4-dab40ee201fb","Type":"ContainerStarted","Data":"efd152bdf42842a8e874fccc8a72147359926a734291adaf54af92d39b3aee56"} Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.314466 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.314789 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.81477473 +0000 UTC m=+163.168689711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.315995 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" event={"ID":"17dcfe6a-c319-4ac4-82a9-407601bf589b","Type":"ContainerStarted","Data":"f7d84e8e44a37ff8c9339be58fb544a3e517e472d8063c85a2db6d2af13da250"} Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.330960 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vkzx6" event={"ID":"f65f8689-77b0-4873-af9d-d82f5ec79c53","Type":"ContainerStarted","Data":"7d21f0ae34668379c5760190fc56fb97f75fa6b50bd978ab4b7d1b594a3ffead"} Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.331011 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vkzx6" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.335969 4945 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-m5fwt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.336014 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" podUID="e76deaf4-7012-4889-87bf-1310e42a6f46" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.336183 4945 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rb2p container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.336233 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" podUID="ad1a76d0-a604-4c89-921d-aed33f60fd01" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.336302 4945 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fdfbw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.336321 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" podUID="87d27326-3c94-424c-af7e-6defb922382d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.336384 4945 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-tjzkw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.336398 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" podUID="76a2fafe-8637-4067-a89b-cf6756ba3e09" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.355332 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.361536 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" podStartSLOduration=134.361516119 podStartE2EDuration="2m14.361516119s" podCreationTimestamp="2025-10-08 15:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:53.35716984 +0000 UTC m=+162.711084761" watchObservedRunningTime="2025-10-08 15:15:53.361516119 +0000 UTC m=+162.715431020" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.388394 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k8w22" podStartSLOduration=133.38837565 podStartE2EDuration="2m13.38837565s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:53.384541854 +0000 UTC m=+162.738456765" watchObservedRunningTime="2025-10-08 15:15:53.38837565 +0000 UTC m=+162.742290551" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.407651 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-m7d4b" podStartSLOduration=133.407629111 podStartE2EDuration="2m13.407629111s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:53.405028996 +0000 UTC m=+162.758943897" watchObservedRunningTime="2025-10-08 15:15:53.407629111 +0000 UTC m=+162.761544012" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.416545 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.418149 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:53.918131633 +0000 UTC m=+163.272046534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.460554 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vkzx6" podStartSLOduration=10.460536794 podStartE2EDuration="10.460536794s" podCreationTimestamp="2025-10-08 15:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:53.458079542 +0000 UTC m=+162.811994433" watchObservedRunningTime="2025-10-08 15:15:53.460536794 +0000 UTC m=+162.814451695" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.520084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.520451 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.02043929 +0000 UTC m=+163.374354191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.527769 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-79mrb" podStartSLOduration=133.527754124 podStartE2EDuration="2m13.527754124s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:53.527455935 +0000 UTC m=+162.881370836" watchObservedRunningTime="2025-10-08 15:15:53.527754124 +0000 UTC m=+162.881669015" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.530046 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-9r6ts" podStartSLOduration=133.53004023 podStartE2EDuration="2m13.53004023s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:53.492181024 +0000 UTC m=+162.846095915" watchObservedRunningTime="2025-10-08 15:15:53.53004023 +0000 UTC m=+162.883955131" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.621731 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.622025 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.122010369 +0000 UTC m=+163.475925270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.722701 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.723077 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.223059584 +0000 UTC m=+163.576974485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.824396 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.824571 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.324545691 +0000 UTC m=+163.678460582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.824777 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.825092 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.325080524 +0000 UTC m=+163.678995425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.890006 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:53 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:53 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:53 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.890084 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.925962 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.926198 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.42616055 +0000 UTC m=+163.780075451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:53 crc kubenswrapper[4945]: I1008 15:15:53.926324 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:53 crc kubenswrapper[4945]: E1008 15:15:53.926652 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.426644421 +0000 UTC m=+163.780559432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.026863 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.027017 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.52699429 +0000 UTC m=+163.880909191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.027150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.027447 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.52744004 +0000 UTC m=+163.881354941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.128143 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.128323 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.628298421 +0000 UTC m=+163.982213322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.128417 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.128801 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.628784484 +0000 UTC m=+163.982699465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.230132 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.230332 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.73031024 +0000 UTC m=+164.084225141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.230496 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.230915 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.730893256 +0000 UTC m=+164.084808217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.331165 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.331354 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.831331855 +0000 UTC m=+164.185246746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.331446 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.331819 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.831803367 +0000 UTC m=+164.185718328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.336346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e","Type":"ContainerStarted","Data":"73ecbdef304dbde990c7e9d96a9a235b74052899c0e81dbb5209eb2a0462e251"} Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.336415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e","Type":"ContainerStarted","Data":"fee6ca39d6f19a595b51f3b6e0529a923f9b9748d120547af705434541581875"} Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.339302 4945 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-bjkt9 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.339342 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" podUID="483f2b9f-6fcd-4569-ab81-a6f265083617" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.381562 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.3815425 podStartE2EDuration="2.3815425s" podCreationTimestamp="2025-10-08 15:15:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:54.380160356 +0000 UTC m=+163.734075267" watchObservedRunningTime="2025-10-08 15:15:54.3815425 +0000 UTC m=+163.735457391" Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.432666 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.433575 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:54.93354981 +0000 UTC m=+164.287464711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.534497 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.534864 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.034852672 +0000 UTC m=+164.388767573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.635714 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.636178 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.136147043 +0000 UTC m=+164.490061954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.636280 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.636626 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.136610985 +0000 UTC m=+164.490525946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.737644 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.738088 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.23805804 +0000 UTC m=+164.591972941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.839602 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.839968 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.339952327 +0000 UTC m=+164.693867228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.889774 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:54 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:54 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:54 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.889845 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.940841 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.941030 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.441002792 +0000 UTC m=+164.794917693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:54 crc kubenswrapper[4945]: I1008 15:15:54.941323 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:54 crc kubenswrapper[4945]: E1008 15:15:54.941619 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.441610477 +0000 UTC m=+164.795525378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.041930 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.042306 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.542288553 +0000 UTC m=+164.896203454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.082768 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.082818 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.084074 4945 patch_prober.go:28] interesting pod/apiserver-76f77b778f-sbkjb container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.084159 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" podUID="8e75c21c-179f-4c78-b1f4-dab40ee201fb" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.099330 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.143320 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.143753 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.643733689 +0000 UTC m=+164.997648640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.244694 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.246126 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.746087177 +0000 UTC m=+165.100002088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.342386 4945 generic.go:334] "Generic (PLEG): container finished" podID="d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e" containerID="73ecbdef304dbde990c7e9d96a9a235b74052899c0e81dbb5209eb2a0462e251" exitCode=0 Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.342500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e","Type":"ContainerDied","Data":"73ecbdef304dbde990c7e9d96a9a235b74052899c0e81dbb5209eb2a0462e251"} Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.346424 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.346840 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.846821874 +0000 UTC m=+165.200736865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.400929 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gx9hj"] Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.401818 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.404131 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.418855 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gx9hj"] Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.447696 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.447856 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.947834459 +0000 UTC m=+165.301749380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.448057 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.448708 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:55.948698461 +0000 UTC m=+165.302613472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.549218 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.549416 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.049391026 +0000 UTC m=+165.403305927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.549472 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.549684 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-utilities\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.549747 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.049735245 +0000 UTC m=+165.403650146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.550052 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sww7c\" (UniqueName: \"kubernetes.io/projected/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-kube-api-access-sww7c\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.550177 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-catalog-content\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.597492 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rrx2t"] Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.598644 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: W1008 15:15:55.603398 4945 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.603440 4945 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.611573 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rrx2t"] Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.650818 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.651038 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.151001996 +0000 UTC m=+165.504916917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.651213 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcnck\" (UniqueName: \"kubernetes.io/projected/5cfb91ef-187d-4876-bf59-d86a23eb0b74-kube-api-access-qcnck\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.651261 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.651286 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-utilities\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.651363 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-utilities\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.651421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sww7c\" (UniqueName: \"kubernetes.io/projected/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-kube-api-access-sww7c\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.651444 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-catalog-content\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.651476 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-catalog-content\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.651588 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.15157903 +0000 UTC m=+165.505493931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.652484 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-catalog-content\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.652531 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-utilities\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.672440 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sww7c\" (UniqueName: \"kubernetes.io/projected/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-kube-api-access-sww7c\") pod \"community-operators-gx9hj\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.716535 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.752594 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.752716 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.252696638 +0000 UTC m=+165.606611549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.753345 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-catalog-content\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.753801 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-catalog-content\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.755183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcnck\" (UniqueName: \"kubernetes.io/projected/5cfb91ef-187d-4876-bf59-d86a23eb0b74-kube-api-access-qcnck\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.755245 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.755274 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-utilities\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.755598 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.25558548 +0000 UTC m=+165.609500381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.755601 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-utilities\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.771654 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcnck\" (UniqueName: \"kubernetes.io/projected/5cfb91ef-187d-4876-bf59-d86a23eb0b74-kube-api-access-qcnck\") pod \"certified-operators-rrx2t\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.800621 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f5zbx"] Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.801919 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.807897 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5zbx"] Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.832159 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.832325 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.833736 4945 patch_prober.go:28] interesting pod/console-f9d7485db-kr5ql container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.833822 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kr5ql" podUID="6df6697f-bec8-4b87-a0ef-7e71c9a0b360" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.857516 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.857774 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.857782 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.357755683 +0000 UTC m=+165.711670584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.857857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-catalog-content\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.858031 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf5m8\" (UniqueName: \"kubernetes.io/projected/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-kube-api-access-vf5m8\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.858098 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.858194 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-utilities\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.859679 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.859740 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.859768 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.359751163 +0000 UTC m=+165.713666104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.860198 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.860242 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.860384 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.860412 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.887647 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:55 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:55 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:55 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.887725 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.911354 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.922422 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.929234 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.929283 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.959923 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.960098 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.46007129 +0000 UTC m=+165.813986191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.960379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-utilities\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.960530 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-catalog-content\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.960788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf5m8\" (UniqueName: \"kubernetes.io/projected/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-kube-api-access-vf5m8\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.960859 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-utilities\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.960912 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:55 crc kubenswrapper[4945]: E1008 15:15:55.961412 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.461399584 +0000 UTC m=+165.815314485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.961820 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-catalog-content\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:55 crc kubenswrapper[4945]: I1008 15:15:55.992929 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf5m8\" (UniqueName: \"kubernetes.io/projected/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-kube-api-access-vf5m8\") pod \"community-operators-f5zbx\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.000519 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rpvxx"] Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.001434 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.022602 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpvxx"] Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.036487 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bjkt9" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.062265 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.062658 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-catalog-content\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.062933 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw7z6\" (UniqueName: \"kubernetes.io/projected/05326826-d64d-44d4-8cf0-1bd591c4fbf2-kube-api-access-qw7z6\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.063048 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-utilities\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.064084 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.564042908 +0000 UTC m=+165.917957879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.125692 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.132772 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.164656 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.164718 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw7z6\" (UniqueName: \"kubernetes.io/projected/05326826-d64d-44d4-8cf0-1bd591c4fbf2-kube-api-access-qw7z6\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.165041 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.665022932 +0000 UTC m=+166.018937833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.165086 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-utilities\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.165216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-catalog-content\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.165514 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-utilities\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.165557 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-catalog-content\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.171981 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-b5mwn" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.180744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw7z6\" (UniqueName: \"kubernetes.io/projected/05326826-d64d-44d4-8cf0-1bd591c4fbf2-kube-api-access-qw7z6\") pod \"certified-operators-rpvxx\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.209904 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.266798 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.266907 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.766887848 +0000 UTC m=+166.120802759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.267387 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.267867 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.767847412 +0000 UTC m=+166.121762313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.362204 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pjg7s" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.369303 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.369401 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.869385929 +0000 UTC m=+166.223300830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.369558 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.369836 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.86982943 +0000 UTC m=+166.223744331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.424713 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5zbx"] Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.471638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.472036 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.972006714 +0000 UTC m=+166.325921645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.472181 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.474668 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:56.9746527 +0000 UTC m=+166.328567601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.487711 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gx9hj"] Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.577057 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.577527 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.07748705 +0000 UTC m=+166.431401951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.578394 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.617251 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.674533 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tjzkw" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.679774 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.680349 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.18033888 +0000 UTC m=+166.534253781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.781434 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.781519 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kube-api-access\") pod \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.781536 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kubelet-dir\") pod \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\" (UID: \"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e\") " Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.781623 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.281598421 +0000 UTC m=+166.635513322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.781659 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e" (UID: "d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.781827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.781911 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.782071 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.282064603 +0000 UTC m=+166.635979504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.790362 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e" (UID: "d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.883168 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.883351 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.383327323 +0000 UTC m=+166.737242224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.883624 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.883726 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.883990 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.38398241 +0000 UTC m=+166.737897311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.885015 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.888181 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rb2p" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.889529 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:56 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:56 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:56 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.889576 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.914841 4945 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/certified-operators-rrx2t" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.915091 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.937976 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fdfbw" Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.985063 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.985235 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.48520146 +0000 UTC m=+166.839116361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:56 crc kubenswrapper[4945]: I1008 15:15:56.985461 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:56 crc kubenswrapper[4945]: E1008 15:15:56.985731 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.485720012 +0000 UTC m=+166.839634913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.087182 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.087903 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.587884185 +0000 UTC m=+166.941799096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.149599 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.150155 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e" containerName="pruner" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.150171 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e" containerName="pruner" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.150293 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e" containerName="pruner" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.150710 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.155637 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.155851 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.159303 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.167639 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.176896 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.189893 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.190311 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.690299205 +0000 UTC m=+167.044214106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.253780 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rrx2t"] Oct 08 15:15:57 crc kubenswrapper[4945]: W1008 15:15:57.268395 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cfb91ef_187d_4876_bf59_d86a23eb0b74.slice/crio-53ea517aaab9c1daf253a4c6b62023a569fcfab7faf3f58908b2476ab2982386 WatchSource:0}: Error finding container 53ea517aaab9c1daf253a4c6b62023a569fcfab7faf3f58908b2476ab2982386: Status 404 returned error can't find the container with id 53ea517aaab9c1daf253a4c6b62023a569fcfab7faf3f58908b2476ab2982386 Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.291566 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.291988 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a54cb337-afd0-4ced-832e-62578051f906-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.292036 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a54cb337-afd0-4ced-832e-62578051f906-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.292202 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.792170171 +0000 UTC m=+167.146085072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.368375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrx2t" event={"ID":"5cfb91ef-187d-4876-bf59-d86a23eb0b74","Type":"ContainerStarted","Data":"53ea517aaab9c1daf253a4c6b62023a569fcfab7faf3f58908b2476ab2982386"} Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.369933 4945 generic.go:334] "Generic (PLEG): container finished" podID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerID="37e05096d57b26f4d3a94d8d0aadbd95dc220cd936b65a231a82aba79cb99b79" exitCode=0 Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.370006 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5zbx" event={"ID":"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7","Type":"ContainerDied","Data":"37e05096d57b26f4d3a94d8d0aadbd95dc220cd936b65a231a82aba79cb99b79"} Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.370071 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5zbx" event={"ID":"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7","Type":"ContainerStarted","Data":"535c3f20bfc49f48a3405a5eca44fa6878168fca6cf61849c71f51fdaf3df991"} Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.371766 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d64cf5f2-c6ff-4bf4-8b53-af33044f9f1e","Type":"ContainerDied","Data":"fee6ca39d6f19a595b51f3b6e0529a923f9b9748d120547af705434541581875"} Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.371804 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fee6ca39d6f19a595b51f3b6e0529a923f9b9748d120547af705434541581875" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.371780 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.372909 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.373493 4945 generic.go:334] "Generic (PLEG): container finished" podID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerID="8c6e0224ce05d4717d3e81960458f58378dbb7774acb671ce1f7d37d69e941d1" exitCode=0 Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.373544 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx9hj" event={"ID":"0ac7a743-3ca9-4a2f-832b-3392d945b7b5","Type":"ContainerDied","Data":"8c6e0224ce05d4717d3e81960458f58378dbb7774acb671ce1f7d37d69e941d1"} Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.373574 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx9hj" event={"ID":"0ac7a743-3ca9-4a2f-832b-3392d945b7b5","Type":"ContainerStarted","Data":"40fb20db8ff7f5ef418098184bc6eb89d57f5c14c61d3175b02c3762f2a0151f"} Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.393018 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a54cb337-afd0-4ced-832e-62578051f906-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.393056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a54cb337-afd0-4ced-832e-62578051f906-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.393083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.393241 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a54cb337-afd0-4ced-832e-62578051f906-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.393402 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.89339092 +0000 UTC m=+167.247305821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.421785 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a54cb337-afd0-4ced-832e-62578051f906-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.446209 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpvxx"] Oct 08 15:15:57 crc kubenswrapper[4945]: W1008 15:15:57.451815 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05326826_d64d_44d4_8cf0_1bd591c4fbf2.slice/crio-c51f37a84a5865d9a71ceeadc7834ea347630069f45486f981e83b8d3b8b0d86 WatchSource:0}: Error finding container c51f37a84a5865d9a71ceeadc7834ea347630069f45486f981e83b8d3b8b0d86: Status 404 returned error can't find the container with id c51f37a84a5865d9a71ceeadc7834ea347630069f45486f981e83b8d3b8b0d86 Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.494019 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.494183 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.994160819 +0000 UTC m=+167.348075730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.494440 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.494737 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:57.994724623 +0000 UTC m=+167.348639534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.578439 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.594878 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.595015 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.094987918 +0000 UTC m=+167.448902819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.595212 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.595573 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.095564093 +0000 UTC m=+167.449478994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.630847 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xrqmk"] Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.631911 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.634076 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.646184 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrqmk"] Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.696407 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.696655 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-catalog-content\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.696708 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-utilities\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.696728 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55hsq\" (UniqueName: \"kubernetes.io/projected/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-kube-api-access-55hsq\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.696874 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.196859175 +0000 UTC m=+167.550774066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.798034 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-catalog-content\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.798382 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-utilities\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.798401 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55hsq\" (UniqueName: \"kubernetes.io/projected/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-kube-api-access-55hsq\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.798440 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.798769 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.298756891 +0000 UTC m=+167.652671792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.799357 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-catalog-content\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.800789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-utilities\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.817519 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.824156 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55hsq\" (UniqueName: \"kubernetes.io/projected/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-kube-api-access-55hsq\") pod \"redhat-marketplace-xrqmk\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.889189 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:57 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:57 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:57 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.889276 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.899089 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:57 crc kubenswrapper[4945]: W1008 15:15:57.899220 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda54cb337_afd0_4ced_832e_62578051f906.slice/crio-28546c9a9d509dbe437446e005f5ffbc3a153bc1c0fe591139b0ece398d68e0f WatchSource:0}: Error finding container 28546c9a9d509dbe437446e005f5ffbc3a153bc1c0fe591139b0ece398d68e0f: Status 404 returned error can't find the container with id 28546c9a9d509dbe437446e005f5ffbc3a153bc1c0fe591139b0ece398d68e0f Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.899318 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.399298104 +0000 UTC m=+167.753213015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.899773 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:57 crc kubenswrapper[4945]: E1008 15:15:57.900153 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.400142225 +0000 UTC m=+167.754057126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:57 crc kubenswrapper[4945]: I1008 15:15:57.951927 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.000650 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.001085 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.501066437 +0000 UTC m=+167.854981338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.005438 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hnmhd"] Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.006561 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.010383 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnmhd"] Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.102123 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-utilities\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.102200 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jgjc\" (UniqueName: \"kubernetes.io/projected/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-kube-api-access-6jgjc\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.102241 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-catalog-content\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.102265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.102793 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.602776399 +0000 UTC m=+167.956691300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.203642 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.203883 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.703850575 +0000 UTC m=+168.057765476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.204141 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-catalog-content\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.204165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.204217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-utilities\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.204280 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jgjc\" (UniqueName: \"kubernetes.io/projected/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-kube-api-access-6jgjc\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.204888 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-catalog-content\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.205103 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.705092686 +0000 UTC m=+168.059007587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.205488 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-utilities\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.219574 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrqmk"] Oct 08 15:15:58 crc kubenswrapper[4945]: W1008 15:15:58.232322 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cf6d252_89f0_46bc_a4aa_5c2867e3baf3.slice/crio-aa3ffd9978ca54d7b316834ca782635ac0820ebac5287a8b763839ee486e6184 WatchSource:0}: Error finding container aa3ffd9978ca54d7b316834ca782635ac0820ebac5287a8b763839ee486e6184: Status 404 returned error can't find the container with id aa3ffd9978ca54d7b316834ca782635ac0820ebac5287a8b763839ee486e6184 Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.244522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jgjc\" (UniqueName: \"kubernetes.io/projected/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-kube-api-access-6jgjc\") pod \"redhat-marketplace-hnmhd\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.306270 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.306759 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.806740587 +0000 UTC m=+168.160655478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.356813 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.384276 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerStarted","Data":"aa3ffd9978ca54d7b316834ca782635ac0820ebac5287a8b763839ee486e6184"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.388659 4945 generic.go:334] "Generic (PLEG): container finished" podID="2ba793aa-5158-46a3-8ea6-1e91389524d1" containerID="2e028ceec119da365ac3a511d19e8c0236278e93a94d7c9f0c81610de0295962" exitCode=0 Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.388714 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" event={"ID":"2ba793aa-5158-46a3-8ea6-1e91389524d1","Type":"ContainerDied","Data":"2e028ceec119da365ac3a511d19e8c0236278e93a94d7c9f0c81610de0295962"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.392199 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" event={"ID":"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7","Type":"ContainerStarted","Data":"717409c15a2f29946833dfff56fc3faa7118e09cceddb3ea46434a65d0186f7f"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.392225 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" event={"ID":"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7","Type":"ContainerStarted","Data":"8f104f2dbf0ee81c1a3299fab791018868ccda8413bb013165fe82d8037f7b79"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.404223 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a54cb337-afd0-4ced-832e-62578051f906","Type":"ContainerStarted","Data":"28546c9a9d509dbe437446e005f5ffbc3a153bc1c0fe591139b0ece398d68e0f"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.407101 4945 generic.go:334] "Generic (PLEG): container finished" podID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerID="899985b52404eb307d0963509f8e01a870abbc983349a9c778342bb538008bb6" exitCode=0 Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.407195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpvxx" event={"ID":"05326826-d64d-44d4-8cf0-1bd591c4fbf2","Type":"ContainerDied","Data":"899985b52404eb307d0963509f8e01a870abbc983349a9c778342bb538008bb6"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.407229 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpvxx" event={"ID":"05326826-d64d-44d4-8cf0-1bd591c4fbf2","Type":"ContainerStarted","Data":"c51f37a84a5865d9a71ceeadc7834ea347630069f45486f981e83b8d3b8b0d86"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.408860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.409185 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:58.909172616 +0000 UTC m=+168.263087517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.435689 4945 generic.go:334] "Generic (PLEG): container finished" podID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerID="390dffb6684ce85858d269f3a21f53acdbca17a88c20fcb80e5a9a056efff58e" exitCode=0 Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.436917 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrx2t" event={"ID":"5cfb91ef-187d-4876-bf59-d86a23eb0b74","Type":"ContainerDied","Data":"390dffb6684ce85858d269f3a21f53acdbca17a88c20fcb80e5a9a056efff58e"} Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.461354 4945 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.510137 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.510275 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.010257702 +0000 UTC m=+168.364172603 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.510393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.511550 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.011542615 +0000 UTC m=+168.365457516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.591358 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnmhd"] Oct 08 15:15:58 crc kubenswrapper[4945]: W1008 15:15:58.597297 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcb19c22_f621_450d_b675_1b2c3a1c4dbc.slice/crio-33b7e97ef85430f8c3983dc5ad6b5d7e1bf5e196070b396f58e9eeb81d3f5f48 WatchSource:0}: Error finding container 33b7e97ef85430f8c3983dc5ad6b5d7e1bf5e196070b396f58e9eeb81d3f5f48: Status 404 returned error can't find the container with id 33b7e97ef85430f8c3983dc5ad6b5d7e1bf5e196070b396f58e9eeb81d3f5f48 Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.599350 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fbbr4"] Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.600652 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.602867 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.608816 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fbbr4"] Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.611263 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.611380 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.111359159 +0000 UTC m=+168.465274060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.611458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.611798 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.11179093 +0000 UTC m=+168.465705821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.712861 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.713062 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.2130372 +0000 UTC m=+168.566952091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.713214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-utilities\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.713242 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.713290 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-catalog-content\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.713382 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzjs8\" (UniqueName: \"kubernetes.io/projected/7cff81a6-1ecf-4615-a620-805f9fa659b0-kube-api-access-nzjs8\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.713517 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.213505962 +0000 UTC m=+168.567420863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.814366 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.814509 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.314485115 +0000 UTC m=+168.668400016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.814666 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-utilities\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.814702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.814756 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-catalog-content\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.814785 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzjs8\" (UniqueName: \"kubernetes.io/projected/7cff81a6-1ecf-4615-a620-805f9fa659b0-kube-api-access-nzjs8\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.815141 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-utilities\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.815413 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.315401369 +0000 UTC m=+168.669316320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.815524 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-catalog-content\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.837728 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzjs8\" (UniqueName: \"kubernetes.io/projected/7cff81a6-1ecf-4615-a620-805f9fa659b0-kube-api-access-nzjs8\") pod \"redhat-operators-fbbr4\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.888531 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:58 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:58 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:58 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.888584 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.916348 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.916495 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.416476634 +0000 UTC m=+168.770391535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.916654 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:58 crc kubenswrapper[4945]: E1008 15:15:58.916936 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.416926466 +0000 UTC m=+168.770841367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:58 crc kubenswrapper[4945]: I1008 15:15:58.922102 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.002138 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-64v6d"] Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.009452 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.012593 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-64v6d"] Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.017360 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:59 crc kubenswrapper[4945]: E1008 15:15:59.017492 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.517468608 +0000 UTC m=+168.871383509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.017735 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:59 crc kubenswrapper[4945]: E1008 15:15:59.018078 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.518065764 +0000 UTC m=+168.871980685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.119432 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:59 crc kubenswrapper[4945]: E1008 15:15:59.119573 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.619545189 +0000 UTC m=+168.973460090 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.119698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t5x8\" (UniqueName: \"kubernetes.io/projected/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-kube-api-access-4t5x8\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.119736 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-utilities\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.119847 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.119940 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-catalog-content\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: E1008 15:15:59.120554 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.620529604 +0000 UTC m=+168.974444505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.221566 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:59 crc kubenswrapper[4945]: E1008 15:15:59.221742 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.721718443 +0000 UTC m=+169.075633364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.221815 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-catalog-content\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.221875 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t5x8\" (UniqueName: \"kubernetes.io/projected/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-kube-api-access-4t5x8\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.221912 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-utilities\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.222401 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-utilities\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.222574 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:59 crc kubenswrapper[4945]: E1008 15:15:59.222902 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 15:15:59.722892092 +0000 UTC m=+169.076806993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-922gb" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.223843 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-catalog-content\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.241425 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t5x8\" (UniqueName: \"kubernetes.io/projected/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-kube-api-access-4t5x8\") pod \"redhat-operators-64v6d\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.304969 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fbbr4"] Oct 08 15:15:59 crc kubenswrapper[4945]: W1008 15:15:59.312016 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cff81a6_1ecf_4615_a620_805f9fa659b0.slice/crio-b552e010444dd001dd3662c45c81fa4445946afdd90fc51b398cc6bec65c166e WatchSource:0}: Error finding container b552e010444dd001dd3662c45c81fa4445946afdd90fc51b398cc6bec65c166e: Status 404 returned error can't find the container with id b552e010444dd001dd3662c45c81fa4445946afdd90fc51b398cc6bec65c166e Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.316208 4945 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-08T15:15:58.461578016Z","Handler":null,"Name":""} Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.324222 4945 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.324276 4945 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.324335 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.327422 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.329959 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.426154 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.428259 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.428285 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.442162 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnmhd" event={"ID":"fcb19c22-f621-450d-b675-1b2c3a1c4dbc","Type":"ContainerStarted","Data":"33b7e97ef85430f8c3983dc5ad6b5d7e1bf5e196070b396f58e9eeb81d3f5f48"} Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.447430 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerStarted","Data":"d5a2e88ad436714c340076f875b1c32c18b76f6a5fdddcf2d201a61e34857675"} Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.448918 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fbbr4" event={"ID":"7cff81a6-1ecf-4615-a620-805f9fa659b0","Type":"ContainerStarted","Data":"b552e010444dd001dd3662c45c81fa4445946afdd90fc51b398cc6bec65c166e"} Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.452215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a54cb337-afd0-4ced-832e-62578051f906","Type":"ContainerStarted","Data":"82634be96e6e2682378d2ef66e927fbe8ced09f54269e66b50bbe5d2efc014b5"} Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.455957 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-922gb\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.468029 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.468004988 podStartE2EDuration="2.468004988s" podCreationTimestamp="2025-10-08 15:15:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:15:59.464342637 +0000 UTC m=+168.818257538" watchObservedRunningTime="2025-10-08 15:15:59.468004988 +0000 UTC m=+168.821919899" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.511476 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-64v6d"] Oct 08 15:15:59 crc kubenswrapper[4945]: W1008 15:15:59.520783 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcdc5718_c74a_4abc_80f7_d7de2bd829d1.slice/crio-f70e30576b9cabe46671f55918afdbdf9bccc460fb9db0bc86b1cddb7b71c0d7 WatchSource:0}: Error finding container f70e30576b9cabe46671f55918afdbdf9bccc460fb9db0bc86b1cddb7b71c0d7: Status 404 returned error can't find the container with id f70e30576b9cabe46671f55918afdbdf9bccc460fb9db0bc86b1cddb7b71c0d7 Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.635458 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.729261 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ba793aa-5158-46a3-8ea6-1e91389524d1-config-volume\") pod \"2ba793aa-5158-46a3-8ea6-1e91389524d1\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.729407 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9ft2\" (UniqueName: \"kubernetes.io/projected/2ba793aa-5158-46a3-8ea6-1e91389524d1-kube-api-access-j9ft2\") pod \"2ba793aa-5158-46a3-8ea6-1e91389524d1\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.729485 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ba793aa-5158-46a3-8ea6-1e91389524d1-secret-volume\") pod \"2ba793aa-5158-46a3-8ea6-1e91389524d1\" (UID: \"2ba793aa-5158-46a3-8ea6-1e91389524d1\") " Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.731878 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ba793aa-5158-46a3-8ea6-1e91389524d1-config-volume" (OuterVolumeSpecName: "config-volume") pod "2ba793aa-5158-46a3-8ea6-1e91389524d1" (UID: "2ba793aa-5158-46a3-8ea6-1e91389524d1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.734878 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba793aa-5158-46a3-8ea6-1e91389524d1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2ba793aa-5158-46a3-8ea6-1e91389524d1" (UID: "2ba793aa-5158-46a3-8ea6-1e91389524d1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.736304 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ba793aa-5158-46a3-8ea6-1e91389524d1-kube-api-access-j9ft2" (OuterVolumeSpecName: "kube-api-access-j9ft2") pod "2ba793aa-5158-46a3-8ea6-1e91389524d1" (UID: "2ba793aa-5158-46a3-8ea6-1e91389524d1"). InnerVolumeSpecName "kube-api-access-j9ft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.752318 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.831186 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ba793aa-5158-46a3-8ea6-1e91389524d1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.831220 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9ft2\" (UniqueName: \"kubernetes.io/projected/2ba793aa-5158-46a3-8ea6-1e91389524d1-kube-api-access-j9ft2\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.831235 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ba793aa-5158-46a3-8ea6-1e91389524d1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.888472 4945 patch_prober.go:28] interesting pod/router-default-5444994796-pr9cp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 15:15:59 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 08 15:15:59 crc kubenswrapper[4945]: [+]process-running ok Oct 08 15:15:59 crc kubenswrapper[4945]: healthz check failed Oct 08 15:15:59 crc kubenswrapper[4945]: I1008 15:15:59.888521 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pr9cp" podUID="b32adce4-0f8a-433e-8643-a9b35e02fd8d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.032191 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.068761 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-922gb"] Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.090678 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.095454 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-sbkjb" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.455840 4945 generic.go:334] "Generic (PLEG): container finished" podID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerID="2cf5d9b998e547999189f9001dcc7ed904fe4d92f8022806cfcb78bbbb67f302" exitCode=0 Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.455918 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnmhd" event={"ID":"fcb19c22-f621-450d-b675-1b2c3a1c4dbc","Type":"ContainerDied","Data":"2cf5d9b998e547999189f9001dcc7ed904fe4d92f8022806cfcb78bbbb67f302"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.458621 4945 generic.go:334] "Generic (PLEG): container finished" podID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerID="d5a2e88ad436714c340076f875b1c32c18b76f6a5fdddcf2d201a61e34857675" exitCode=0 Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.458803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerDied","Data":"d5a2e88ad436714c340076f875b1c32c18b76f6a5fdddcf2d201a61e34857675"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.460230 4945 generic.go:334] "Generic (PLEG): container finished" podID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerID="95a6a9f989163672933e0b2d62e1e821271c1aa464d8ae723e0bfdc9b9e0d935" exitCode=0 Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.460302 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fbbr4" event={"ID":"7cff81a6-1ecf-4615-a620-805f9fa659b0","Type":"ContainerDied","Data":"95a6a9f989163672933e0b2d62e1e821271c1aa464d8ae723e0bfdc9b9e0d935"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.463060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" event={"ID":"2ba793aa-5158-46a3-8ea6-1e91389524d1","Type":"ContainerDied","Data":"25990d6267527ec1414b1c6a15fb1738d1e4cb2efa57c0a1f973e1617fc5e56f"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.463082 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25990d6267527ec1414b1c6a15fb1738d1e4cb2efa57c0a1f973e1617fc5e56f" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.463179 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.465936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" event={"ID":"a5e59ddd-f9da-45e2-93a8-dc8d3b45d8e7","Type":"ContainerStarted","Data":"ae7f04846d92282203531453c2211a76fc353303b572fb51ee80028cf4dd8d47"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.468793 4945 generic.go:334] "Generic (PLEG): container finished" podID="a54cb337-afd0-4ced-832e-62578051f906" containerID="82634be96e6e2682378d2ef66e927fbe8ced09f54269e66b50bbe5d2efc014b5" exitCode=0 Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.468886 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a54cb337-afd0-4ced-832e-62578051f906","Type":"ContainerDied","Data":"82634be96e6e2682378d2ef66e927fbe8ced09f54269e66b50bbe5d2efc014b5"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.470460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" event={"ID":"5fec4a94-36ab-4892-a13c-35a21108d2ba","Type":"ContainerStarted","Data":"78b07e3004bdc8823f90dd1cbe6f71ffe7420ddc55ddaaea66c54d420abb67c6"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.471309 4945 generic.go:334] "Generic (PLEG): container finished" podID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerID="520954aa2a83fb1a9bcb6a80baf3e60d552de350615a010c9ab8653fb42f178b" exitCode=0 Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.471410 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64v6d" event={"ID":"bcdc5718-c74a-4abc-80f7-d7de2bd829d1","Type":"ContainerDied","Data":"520954aa2a83fb1a9bcb6a80baf3e60d552de350615a010c9ab8653fb42f178b"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.471448 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64v6d" event={"ID":"bcdc5718-c74a-4abc-80f7-d7de2bd829d1","Type":"ContainerStarted","Data":"f70e30576b9cabe46671f55918afdbdf9bccc460fb9db0bc86b1cddb7b71c0d7"} Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.613537 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6trg2" podStartSLOduration=17.613517256 podStartE2EDuration="17.613517256s" podCreationTimestamp="2025-10-08 15:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:16:00.56967655 +0000 UTC m=+169.923591451" watchObservedRunningTime="2025-10-08 15:16:00.613517256 +0000 UTC m=+169.967432157" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.887812 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:16:00 crc kubenswrapper[4945]: I1008 15:16:00.889905 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-pr9cp" Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.479338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" event={"ID":"5fec4a94-36ab-4892-a13c-35a21108d2ba","Type":"ContainerStarted","Data":"25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff"} Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.578008 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vkzx6" Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.596301 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" podStartSLOduration=141.596277146 podStartE2EDuration="2m21.596277146s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:16:01.528599505 +0000 UTC m=+170.882514406" watchObservedRunningTime="2025-10-08 15:16:01.596277146 +0000 UTC m=+170.950192047" Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.704954 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.758638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a54cb337-afd0-4ced-832e-62578051f906-kube-api-access\") pod \"a54cb337-afd0-4ced-832e-62578051f906\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.758728 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a54cb337-afd0-4ced-832e-62578051f906-kubelet-dir\") pod \"a54cb337-afd0-4ced-832e-62578051f906\" (UID: \"a54cb337-afd0-4ced-832e-62578051f906\") " Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.760280 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a54cb337-afd0-4ced-832e-62578051f906-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a54cb337-afd0-4ced-832e-62578051f906" (UID: "a54cb337-afd0-4ced-832e-62578051f906"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.760721 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a54cb337-afd0-4ced-832e-62578051f906-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.764936 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54cb337-afd0-4ced-832e-62578051f906-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a54cb337-afd0-4ced-832e-62578051f906" (UID: "a54cb337-afd0-4ced-832e-62578051f906"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:16:01 crc kubenswrapper[4945]: I1008 15:16:01.861491 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a54cb337-afd0-4ced-832e-62578051f906-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 15:16:02 crc kubenswrapper[4945]: I1008 15:16:02.486166 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a54cb337-afd0-4ced-832e-62578051f906","Type":"ContainerDied","Data":"28546c9a9d509dbe437446e005f5ffbc3a153bc1c0fe591139b0ece398d68e0f"} Oct 08 15:16:02 crc kubenswrapper[4945]: I1008 15:16:02.486211 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28546c9a9d509dbe437446e005f5ffbc3a153bc1c0fe591139b0ece398d68e0f" Oct 08 15:16:02 crc kubenswrapper[4945]: I1008 15:16:02.486234 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 15:16:02 crc kubenswrapper[4945]: I1008 15:16:02.486331 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:16:02 crc kubenswrapper[4945]: I1008 15:16:02.570069 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:16:02 crc kubenswrapper[4945]: I1008 15:16:02.573756 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0-metrics-certs\") pod \"network-metrics-daemon-zj7s7\" (UID: \"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0\") " pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:16:02 crc kubenswrapper[4945]: I1008 15:16:02.840731 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zj7s7" Oct 08 15:16:03 crc kubenswrapper[4945]: I1008 15:16:03.049853 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zj7s7"] Oct 08 15:16:03 crc kubenswrapper[4945]: W1008 15:16:03.065350 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e117c8d_02e9_45ae_8cf6_80a13b3bf4d0.slice/crio-6552b238a58fb6c869c2914a4ee1d581bdfe52fd6afd59e74af151ae50e10530 WatchSource:0}: Error finding container 6552b238a58fb6c869c2914a4ee1d581bdfe52fd6afd59e74af151ae50e10530: Status 404 returned error can't find the container with id 6552b238a58fb6c869c2914a4ee1d581bdfe52fd6afd59e74af151ae50e10530 Oct 08 15:16:03 crc kubenswrapper[4945]: I1008 15:16:03.506158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" event={"ID":"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0","Type":"ContainerStarted","Data":"a08aa1292a9ded56918498654d8f50314d1329049851946c471a69ea6c547a24"} Oct 08 15:16:03 crc kubenswrapper[4945]: I1008 15:16:03.506494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" event={"ID":"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0","Type":"ContainerStarted","Data":"6552b238a58fb6c869c2914a4ee1d581bdfe52fd6afd59e74af151ae50e10530"} Oct 08 15:16:04 crc kubenswrapper[4945]: I1008 15:16:04.513836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zj7s7" event={"ID":"6e117c8d-02e9-45ae-8cf6-80a13b3bf4d0","Type":"ContainerStarted","Data":"65fea234d4b0d9aaf405d230cfd023be0695e6434298a94e585e1b9fab743096"} Oct 08 15:16:05 crc kubenswrapper[4945]: I1008 15:16:05.837075 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:16:05 crc kubenswrapper[4945]: I1008 15:16:05.841268 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:16:05 crc kubenswrapper[4945]: I1008 15:16:05.853214 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zj7s7" podStartSLOduration=145.853195301 podStartE2EDuration="2m25.853195301s" podCreationTimestamp="2025-10-08 15:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:16:04.539297325 +0000 UTC m=+173.893212226" watchObservedRunningTime="2025-10-08 15:16:05.853195301 +0000 UTC m=+175.207110202" Oct 08 15:16:05 crc kubenswrapper[4945]: I1008 15:16:05.857896 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:05 crc kubenswrapper[4945]: I1008 15:16:05.857897 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:05 crc kubenswrapper[4945]: I1008 15:16:05.857955 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:05 crc kubenswrapper[4945]: I1008 15:16:05.857993 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.858892 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.859020 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.859444 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.859533 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.859797 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.860509 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.860536 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.860797 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"d4a3168436dfbe3143ff00dda0bc8d66f8f846030d827a0ac108c0318e480069"} pod="openshift-console/downloads-7954f5f757-hb8lk" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 08 15:16:15 crc kubenswrapper[4945]: I1008 15:16:15.860899 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" containerID="cri-o://d4a3168436dfbe3143ff00dda0bc8d66f8f846030d827a0ac108c0318e480069" gracePeriod=2 Oct 08 15:16:19 crc kubenswrapper[4945]: I1008 15:16:19.184779 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:16:19 crc kubenswrapper[4945]: I1008 15:16:19.185517 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:16:19 crc kubenswrapper[4945]: I1008 15:16:19.610717 4945 generic.go:334] "Generic (PLEG): container finished" podID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerID="d4a3168436dfbe3143ff00dda0bc8d66f8f846030d827a0ac108c0318e480069" exitCode=0 Oct 08 15:16:19 crc kubenswrapper[4945]: I1008 15:16:19.610764 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hb8lk" event={"ID":"54eb5003-d347-47aa-b14a-391d1bd74bdb","Type":"ContainerDied","Data":"d4a3168436dfbe3143ff00dda0bc8d66f8f846030d827a0ac108c0318e480069"} Oct 08 15:16:19 crc kubenswrapper[4945]: I1008 15:16:19.759586 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:16:25 crc kubenswrapper[4945]: I1008 15:16:25.859699 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:25 crc kubenswrapper[4945]: I1008 15:16:25.860323 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:26 crc kubenswrapper[4945]: I1008 15:16:26.946014 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b6msk" Oct 08 15:16:35 crc kubenswrapper[4945]: I1008 15:16:35.859246 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:35 crc kubenswrapper[4945]: I1008 15:16:35.859956 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:44 crc kubenswrapper[4945]: E1008 15:16:44.579846 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 15:16:44 crc kubenswrapper[4945]: E1008 15:16:44.580410 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4t5x8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-64v6d_openshift-marketplace(bcdc5718-c74a-4abc-80f7-d7de2bd829d1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:16:44 crc kubenswrapper[4945]: E1008 15:16:44.581584 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-64v6d" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" Oct 08 15:16:45 crc kubenswrapper[4945]: I1008 15:16:45.859152 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:45 crc kubenswrapper[4945]: I1008 15:16:45.859236 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:49 crc kubenswrapper[4945]: I1008 15:16:49.184192 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:16:49 crc kubenswrapper[4945]: I1008 15:16:49.184740 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:16:49 crc kubenswrapper[4945]: I1008 15:16:49.184832 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:16:49 crc kubenswrapper[4945]: I1008 15:16:49.185790 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:16:49 crc kubenswrapper[4945]: I1008 15:16:49.185929 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341" gracePeriod=600 Oct 08 15:16:52 crc kubenswrapper[4945]: E1008 15:16:52.837322 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-64v6d" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" Oct 08 15:16:54 crc kubenswrapper[4945]: E1008 15:16:54.731180 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:fc6d1468707e4bcc767e25ba90e295828fee37cd04f9ceaa879288e8fb4d2d84: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:fc6d1468707e4bcc767e25ba90e295828fee37cd04f9ceaa879288e8fb4d2d84\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 15:16:54 crc kubenswrapper[4945]: E1008 15:16:54.731587 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6jgjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hnmhd_openshift-marketplace(fcb19c22-f621-450d-b675-1b2c3a1c4dbc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:fc6d1468707e4bcc767e25ba90e295828fee37cd04f9ceaa879288e8fb4d2d84: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:fc6d1468707e4bcc767e25ba90e295828fee37cd04f9ceaa879288e8fb4d2d84\": context canceled" logger="UnhandledError" Oct 08 15:16:54 crc kubenswrapper[4945]: E1008 15:16:54.732755 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:fc6d1468707e4bcc767e25ba90e295828fee37cd04f9ceaa879288e8fb4d2d84: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:fc6d1468707e4bcc767e25ba90e295828fee37cd04f9ceaa879288e8fb4d2d84\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-hnmhd" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" Oct 08 15:16:55 crc kubenswrapper[4945]: I1008 15:16:55.858367 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:16:55 crc kubenswrapper[4945]: I1008 15:16:55.858448 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.342187 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.342376 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nzjs8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fbbr4_openshift-marketplace(7cff81a6-1ecf-4615-a620-805f9fa659b0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.343683 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fbbr4" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.872136 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hnmhd" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.896635 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.896825 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vf5m8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-f5zbx_openshift-marketplace(2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.898737 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-f5zbx" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.931033 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.931314 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sww7c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gx9hj_openshift-marketplace(0ac7a743-3ca9-4a2f-832b-3392d945b7b5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:16:57 crc kubenswrapper[4945]: E1008 15:16:57.932554 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gx9hj" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" Oct 08 15:16:58 crc kubenswrapper[4945]: I1008 15:16:58.846353 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341" exitCode=0 Oct 08 15:16:58 crc kubenswrapper[4945]: I1008 15:16:58.846458 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341"} Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.478287 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-f5zbx" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.536315 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.536529 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qw7z6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rpvxx_openshift-marketplace(05326826-d64d-44d4-8cf0-1bd591c4fbf2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.538083 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rpvxx" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.561272 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.561446 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55hsq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xrqmk_openshift-marketplace(2cf6d252-89f0-46bc-a4aa-5c2867e3baf3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.562677 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xrqmk" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.595292 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.595818 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qcnck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rrx2t_openshift-marketplace(5cfb91ef-187d-4876-bf59-d86a23eb0b74): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.596932 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rrx2t" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" Oct 08 15:17:00 crc kubenswrapper[4945]: I1008 15:17:00.857048 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"6dba648876d9e856a999caed1abc8004128f384ec2a39cd78a8487b03680a491"} Oct 08 15:17:00 crc kubenswrapper[4945]: I1008 15:17:00.862928 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hb8lk" event={"ID":"54eb5003-d347-47aa-b14a-391d1bd74bdb","Type":"ContainerStarted","Data":"71c40c3f92e177e35b76ee9104191017f1faa05d545354cf03ed4c4ef394ae1a"} Oct 08 15:17:00 crc kubenswrapper[4945]: I1008 15:17:00.863469 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:17:00 crc kubenswrapper[4945]: I1008 15:17:00.863512 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.864930 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rrx2t" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.865543 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xrqmk" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" Oct 08 15:17:00 crc kubenswrapper[4945]: E1008 15:17:00.865602 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rpvxx" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" Oct 08 15:17:01 crc kubenswrapper[4945]: I1008 15:17:01.867793 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:17:01 crc kubenswrapper[4945]: I1008 15:17:01.868118 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:17:01 crc kubenswrapper[4945]: I1008 15:17:01.868160 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:17:02 crc kubenswrapper[4945]: I1008 15:17:02.872070 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:17:02 crc kubenswrapper[4945]: I1008 15:17:02.872386 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:17:05 crc kubenswrapper[4945]: I1008 15:17:05.858313 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:17:05 crc kubenswrapper[4945]: I1008 15:17:05.858844 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:17:05 crc kubenswrapper[4945]: I1008 15:17:05.858334 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-hb8lk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 08 15:17:05 crc kubenswrapper[4945]: I1008 15:17:05.858949 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hb8lk" podUID="54eb5003-d347-47aa-b14a-391d1bd74bdb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 08 15:17:10 crc kubenswrapper[4945]: I1008 15:17:10.923248 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64v6d" event={"ID":"bcdc5718-c74a-4abc-80f7-d7de2bd829d1","Type":"ContainerStarted","Data":"5fb9b94daeaa462fd3c4a5e750576d711a4f878418242bffe0aa1bdb55ab4b3e"} Oct 08 15:17:11 crc kubenswrapper[4945]: I1008 15:17:11.930083 4945 generic.go:334] "Generic (PLEG): container finished" podID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerID="11fa19a09828a5d4272fa5bb0f742e5782770a63eddc8cf146aef739df92fc82" exitCode=0 Oct 08 15:17:11 crc kubenswrapper[4945]: I1008 15:17:11.930172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx9hj" event={"ID":"0ac7a743-3ca9-4a2f-832b-3392d945b7b5","Type":"ContainerDied","Data":"11fa19a09828a5d4272fa5bb0f742e5782770a63eddc8cf146aef739df92fc82"} Oct 08 15:17:11 crc kubenswrapper[4945]: I1008 15:17:11.934764 4945 generic.go:334] "Generic (PLEG): container finished" podID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerID="5fb9b94daeaa462fd3c4a5e750576d711a4f878418242bffe0aa1bdb55ab4b3e" exitCode=0 Oct 08 15:17:11 crc kubenswrapper[4945]: I1008 15:17:11.934912 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64v6d" event={"ID":"bcdc5718-c74a-4abc-80f7-d7de2bd829d1","Type":"ContainerDied","Data":"5fb9b94daeaa462fd3c4a5e750576d711a4f878418242bffe0aa1bdb55ab4b3e"} Oct 08 15:17:11 crc kubenswrapper[4945]: I1008 15:17:11.945275 4945 generic.go:334] "Generic (PLEG): container finished" podID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerID="ba661a525097f0de3a7e2f440798c0c134ea3a7f30b3a4245137b182644a6de3" exitCode=0 Oct 08 15:17:11 crc kubenswrapper[4945]: I1008 15:17:11.945344 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnmhd" event={"ID":"fcb19c22-f621-450d-b675-1b2c3a1c4dbc","Type":"ContainerDied","Data":"ba661a525097f0de3a7e2f440798c0c134ea3a7f30b3a4245137b182644a6de3"} Oct 08 15:17:13 crc kubenswrapper[4945]: I1008 15:17:13.958150 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnmhd" event={"ID":"fcb19c22-f621-450d-b675-1b2c3a1c4dbc","Type":"ContainerStarted","Data":"74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8"} Oct 08 15:17:13 crc kubenswrapper[4945]: I1008 15:17:13.959640 4945 generic.go:334] "Generic (PLEG): container finished" podID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerID="91627156ec90af11146b7ac4df857042389fe06650fb089e2b2c12a8b15fee72" exitCode=0 Oct 08 15:17:13 crc kubenswrapper[4945]: I1008 15:17:13.959703 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5zbx" event={"ID":"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7","Type":"ContainerDied","Data":"91627156ec90af11146b7ac4df857042389fe06650fb089e2b2c12a8b15fee72"} Oct 08 15:17:13 crc kubenswrapper[4945]: I1008 15:17:13.964365 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx9hj" event={"ID":"0ac7a743-3ca9-4a2f-832b-3392d945b7b5","Type":"ContainerStarted","Data":"a593a009cca0af16b9859f8bb0c7a75a9caa8eced37d04d5d69dc502ebb28800"} Oct 08 15:17:13 crc kubenswrapper[4945]: I1008 15:17:13.971794 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64v6d" event={"ID":"bcdc5718-c74a-4abc-80f7-d7de2bd829d1","Type":"ContainerStarted","Data":"61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080"} Oct 08 15:17:13 crc kubenswrapper[4945]: I1008 15:17:13.996943 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hnmhd" podStartSLOduration=4.31994683 podStartE2EDuration="1m16.996922986s" podCreationTimestamp="2025-10-08 15:15:57 +0000 UTC" firstStartedPulling="2025-10-08 15:16:00.457467506 +0000 UTC m=+169.811382407" lastFinishedPulling="2025-10-08 15:17:13.134443642 +0000 UTC m=+242.488358563" observedRunningTime="2025-10-08 15:17:13.977795828 +0000 UTC m=+243.331710739" watchObservedRunningTime="2025-10-08 15:17:13.996922986 +0000 UTC m=+243.350837897" Oct 08 15:17:13 crc kubenswrapper[4945]: I1008 15:17:13.997579 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-64v6d" podStartSLOduration=3.210764222 podStartE2EDuration="1m15.997572962s" podCreationTimestamp="2025-10-08 15:15:58 +0000 UTC" firstStartedPulling="2025-10-08 15:16:00.473006095 +0000 UTC m=+169.826920996" lastFinishedPulling="2025-10-08 15:17:13.259814835 +0000 UTC m=+242.613729736" observedRunningTime="2025-10-08 15:17:13.994060165 +0000 UTC m=+243.347975096" watchObservedRunningTime="2025-10-08 15:17:13.997572962 +0000 UTC m=+243.351487863" Oct 08 15:17:14 crc kubenswrapper[4945]: I1008 15:17:14.020640 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gx9hj" podStartSLOduration=4.451617688 podStartE2EDuration="1m19.020620879s" podCreationTimestamp="2025-10-08 15:15:55 +0000 UTC" firstStartedPulling="2025-10-08 15:15:58.442123009 +0000 UTC m=+167.796037910" lastFinishedPulling="2025-10-08 15:17:13.01112621 +0000 UTC m=+242.365041101" observedRunningTime="2025-10-08 15:17:14.017028729 +0000 UTC m=+243.370943630" watchObservedRunningTime="2025-10-08 15:17:14.020620879 +0000 UTC m=+243.374535780" Oct 08 15:17:14 crc kubenswrapper[4945]: I1008 15:17:14.980764 4945 generic.go:334] "Generic (PLEG): container finished" podID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerID="35b7762f9fd5cd43660f4ad32a3fa17327a43685177d828a2cdddb754f658a2a" exitCode=0 Oct 08 15:17:14 crc kubenswrapper[4945]: I1008 15:17:14.980804 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fbbr4" event={"ID":"7cff81a6-1ecf-4615-a620-805f9fa659b0","Type":"ContainerDied","Data":"35b7762f9fd5cd43660f4ad32a3fa17327a43685177d828a2cdddb754f658a2a"} Oct 08 15:17:15 crc kubenswrapper[4945]: I1008 15:17:15.717706 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:17:15 crc kubenswrapper[4945]: I1008 15:17:15.717753 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:17:15 crc kubenswrapper[4945]: I1008 15:17:15.872741 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-hb8lk" Oct 08 15:17:17 crc kubenswrapper[4945]: I1008 15:17:17.033214 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gx9hj" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="registry-server" probeResult="failure" output=< Oct 08 15:17:17 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 15:17:17 crc kubenswrapper[4945]: > Oct 08 15:17:18 crc kubenswrapper[4945]: I1008 15:17:18.357831 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:17:18 crc kubenswrapper[4945]: I1008 15:17:18.357899 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:17:18 crc kubenswrapper[4945]: I1008 15:17:18.646192 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:17:19 crc kubenswrapper[4945]: I1008 15:17:19.069380 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:17:19 crc kubenswrapper[4945]: I1008 15:17:19.330434 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:17:19 crc kubenswrapper[4945]: I1008 15:17:19.330520 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:17:19 crc kubenswrapper[4945]: I1008 15:17:19.404489 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:17:20 crc kubenswrapper[4945]: I1008 15:17:20.060290 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:17:20 crc kubenswrapper[4945]: I1008 15:17:20.468779 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnmhd"] Oct 08 15:17:21 crc kubenswrapper[4945]: I1008 15:17:21.016554 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hnmhd" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="registry-server" containerID="cri-o://74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8" gracePeriod=2 Oct 08 15:17:22 crc kubenswrapper[4945]: I1008 15:17:22.036551 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5zbx" event={"ID":"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7","Type":"ContainerStarted","Data":"423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729"} Oct 08 15:17:22 crc kubenswrapper[4945]: I1008 15:17:22.273622 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-64v6d"] Oct 08 15:17:22 crc kubenswrapper[4945]: I1008 15:17:22.274271 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-64v6d" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="registry-server" containerID="cri-o://61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" gracePeriod=2 Oct 08 15:17:23 crc kubenswrapper[4945]: I1008 15:17:23.036315 4945 generic.go:334] "Generic (PLEG): container finished" podID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerID="74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8" exitCode=0 Oct 08 15:17:23 crc kubenswrapper[4945]: I1008 15:17:23.036562 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnmhd" event={"ID":"fcb19c22-f621-450d-b675-1b2c3a1c4dbc","Type":"ContainerDied","Data":"74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8"} Oct 08 15:17:23 crc kubenswrapper[4945]: I1008 15:17:23.061497 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f5zbx" podStartSLOduration=4.769363379 podStartE2EDuration="1m28.061476661s" podCreationTimestamp="2025-10-08 15:15:55 +0000 UTC" firstStartedPulling="2025-10-08 15:15:57.37255977 +0000 UTC m=+166.726474671" lastFinishedPulling="2025-10-08 15:17:20.664673042 +0000 UTC m=+250.018587953" observedRunningTime="2025-10-08 15:17:23.056428115 +0000 UTC m=+252.410343016" watchObservedRunningTime="2025-10-08 15:17:23.061476661 +0000 UTC m=+252.415391562" Oct 08 15:17:24 crc kubenswrapper[4945]: I1008 15:17:24.042427 4945 generic.go:334] "Generic (PLEG): container finished" podID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" exitCode=0 Oct 08 15:17:24 crc kubenswrapper[4945]: I1008 15:17:24.042486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64v6d" event={"ID":"bcdc5718-c74a-4abc-80f7-d7de2bd829d1","Type":"ContainerDied","Data":"61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080"} Oct 08 15:17:25 crc kubenswrapper[4945]: I1008 15:17:25.760336 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:17:25 crc kubenswrapper[4945]: I1008 15:17:25.818132 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:17:26 crc kubenswrapper[4945]: I1008 15:17:26.133427 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:17:26 crc kubenswrapper[4945]: I1008 15:17:26.134633 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:17:26 crc kubenswrapper[4945]: I1008 15:17:26.198269 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:17:27 crc kubenswrapper[4945]: I1008 15:17:27.103236 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:17:28 crc kubenswrapper[4945]: E1008 15:17:28.358272 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8 is running failed: container process not found" containerID="74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:28 crc kubenswrapper[4945]: E1008 15:17:28.359184 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8 is running failed: container process not found" containerID="74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:28 crc kubenswrapper[4945]: E1008 15:17:28.359904 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8 is running failed: container process not found" containerID="74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:28 crc kubenswrapper[4945]: E1008 15:17:28.360045 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-hnmhd" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="registry-server" Oct 08 15:17:29 crc kubenswrapper[4945]: E1008 15:17:29.332236 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:29 crc kubenswrapper[4945]: E1008 15:17:29.333605 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:29 crc kubenswrapper[4945]: E1008 15:17:29.334150 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:29 crc kubenswrapper[4945]: E1008 15:17:29.334252 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-64v6d" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="registry-server" Oct 08 15:17:30 crc kubenswrapper[4945]: I1008 15:17:30.067541 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5zbx"] Oct 08 15:17:30 crc kubenswrapper[4945]: I1008 15:17:30.077842 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f5zbx" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="registry-server" containerID="cri-o://423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729" gracePeriod=2 Oct 08 15:17:32 crc kubenswrapper[4945]: I1008 15:17:32.094690 4945 generic.go:334] "Generic (PLEG): container finished" podID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerID="423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729" exitCode=0 Oct 08 15:17:32 crc kubenswrapper[4945]: I1008 15:17:32.095143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5zbx" event={"ID":"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7","Type":"ContainerDied","Data":"423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729"} Oct 08 15:17:36 crc kubenswrapper[4945]: E1008 15:17:36.134574 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729 is running failed: container process not found" containerID="423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:36 crc kubenswrapper[4945]: E1008 15:17:36.135468 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729 is running failed: container process not found" containerID="423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:36 crc kubenswrapper[4945]: E1008 15:17:36.135939 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729 is running failed: container process not found" containerID="423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:36 crc kubenswrapper[4945]: E1008 15:17:36.136001 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-f5zbx" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="registry-server" Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.563511 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.685918 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-catalog-content\") pod \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.686191 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jgjc\" (UniqueName: \"kubernetes.io/projected/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-kube-api-access-6jgjc\") pod \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.686280 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-utilities\") pod \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\" (UID: \"fcb19c22-f621-450d-b675-1b2c3a1c4dbc\") " Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.687661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-utilities" (OuterVolumeSpecName: "utilities") pod "fcb19c22-f621-450d-b675-1b2c3a1c4dbc" (UID: "fcb19c22-f621-450d-b675-1b2c3a1c4dbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.695309 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-kube-api-access-6jgjc" (OuterVolumeSpecName: "kube-api-access-6jgjc") pod "fcb19c22-f621-450d-b675-1b2c3a1c4dbc" (UID: "fcb19c22-f621-450d-b675-1b2c3a1c4dbc"). InnerVolumeSpecName "kube-api-access-6jgjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.707776 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcb19c22-f621-450d-b675-1b2c3a1c4dbc" (UID: "fcb19c22-f621-450d-b675-1b2c3a1c4dbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.788915 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.788992 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jgjc\" (UniqueName: \"kubernetes.io/projected/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-kube-api-access-6jgjc\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:37 crc kubenswrapper[4945]: I1008 15:17:37.789020 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb19c22-f621-450d-b675-1b2c3a1c4dbc-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:38 crc kubenswrapper[4945]: I1008 15:17:38.138694 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnmhd" event={"ID":"fcb19c22-f621-450d-b675-1b2c3a1c4dbc","Type":"ContainerDied","Data":"33b7e97ef85430f8c3983dc5ad6b5d7e1bf5e196070b396f58e9eeb81d3f5f48"} Oct 08 15:17:38 crc kubenswrapper[4945]: I1008 15:17:38.138754 4945 scope.go:117] "RemoveContainer" containerID="74038318a4de823d3743720237faad82d1b5cdad2a106161c0861c3309a212b8" Oct 08 15:17:38 crc kubenswrapper[4945]: I1008 15:17:38.138818 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnmhd" Oct 08 15:17:38 crc kubenswrapper[4945]: I1008 15:17:38.163955 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnmhd"] Oct 08 15:17:38 crc kubenswrapper[4945]: I1008 15:17:38.168731 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnmhd"] Oct 08 15:17:39 crc kubenswrapper[4945]: E1008 15:17:39.331088 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:39 crc kubenswrapper[4945]: E1008 15:17:39.331679 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:39 crc kubenswrapper[4945]: E1008 15:17:39.332034 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:17:39 crc kubenswrapper[4945]: E1008 15:17:39.332074 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-64v6d" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="registry-server" Oct 08 15:17:40 crc kubenswrapper[4945]: I1008 15:17:40.036241 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" path="/var/lib/kubelet/pods/fcb19c22-f621-450d-b675-1b2c3a1c4dbc/volumes" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.177776 4945 scope.go:117] "RemoveContainer" containerID="ba661a525097f0de3a7e2f440798c0c134ea3a7f30b3a4245137b182644a6de3" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.229356 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.235472 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.339256 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-catalog-content\") pod \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.339310 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-utilities\") pod \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.339424 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf5m8\" (UniqueName: \"kubernetes.io/projected/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-kube-api-access-vf5m8\") pod \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.339457 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t5x8\" (UniqueName: \"kubernetes.io/projected/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-kube-api-access-4t5x8\") pod \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.339532 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-utilities\") pod \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\" (UID: \"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7\") " Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.339555 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-catalog-content\") pod \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\" (UID: \"bcdc5718-c74a-4abc-80f7-d7de2bd829d1\") " Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.340204 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-utilities" (OuterVolumeSpecName: "utilities") pod "bcdc5718-c74a-4abc-80f7-d7de2bd829d1" (UID: "bcdc5718-c74a-4abc-80f7-d7de2bd829d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.340533 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-utilities" (OuterVolumeSpecName: "utilities") pod "2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" (UID: "2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.349337 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-kube-api-access-4t5x8" (OuterVolumeSpecName: "kube-api-access-4t5x8") pod "bcdc5718-c74a-4abc-80f7-d7de2bd829d1" (UID: "bcdc5718-c74a-4abc-80f7-d7de2bd829d1"). InnerVolumeSpecName "kube-api-access-4t5x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.354364 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-kube-api-access-vf5m8" (OuterVolumeSpecName: "kube-api-access-vf5m8") pod "2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" (UID: "2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7"). InnerVolumeSpecName "kube-api-access-vf5m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.418603 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" (UID: "2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.428960 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcdc5718-c74a-4abc-80f7-d7de2bd829d1" (UID: "bcdc5718-c74a-4abc-80f7-d7de2bd829d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.441197 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf5m8\" (UniqueName: \"kubernetes.io/projected/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-kube-api-access-vf5m8\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.441241 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t5x8\" (UniqueName: \"kubernetes.io/projected/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-kube-api-access-4t5x8\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.441259 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.441277 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.441294 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:41 crc kubenswrapper[4945]: I1008 15:17:41.441310 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcdc5718-c74a-4abc-80f7-d7de2bd829d1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.168553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64v6d" event={"ID":"bcdc5718-c74a-4abc-80f7-d7de2bd829d1","Type":"ContainerDied","Data":"f70e30576b9cabe46671f55918afdbdf9bccc460fb9db0bc86b1cddb7b71c0d7"} Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.168900 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64v6d" Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.177392 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5zbx" event={"ID":"2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7","Type":"ContainerDied","Data":"535c3f20bfc49f48a3405a5eca44fa6878168fca6cf61849c71f51fdaf3df991"} Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.177494 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5zbx" Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.189912 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-64v6d"] Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.200688 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-64v6d"] Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.208669 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5zbx"] Oct 08 15:17:42 crc kubenswrapper[4945]: I1008 15:17:42.212615 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f5zbx"] Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.877278 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.879298 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.978152 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.978235 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.978287 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.980457 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.980658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.981175 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.991055 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 15:17:43 crc kubenswrapper[4945]: I1008 15:17:43.995804 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:17:44 crc kubenswrapper[4945]: I1008 15:17:44.004601 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:17:44 crc kubenswrapper[4945]: I1008 15:17:44.004749 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:17:44 crc kubenswrapper[4945]: I1008 15:17:44.032950 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" path="/var/lib/kubelet/pods/2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7/volumes" Oct 08 15:17:44 crc kubenswrapper[4945]: I1008 15:17:44.034090 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" path="/var/lib/kubelet/pods/bcdc5718-c74a-4abc-80f7-d7de2bd829d1/volumes" Oct 08 15:17:44 crc kubenswrapper[4945]: I1008 15:17:44.172482 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 15:17:44 crc kubenswrapper[4945]: I1008 15:17:44.172489 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 15:17:44 crc kubenswrapper[4945]: I1008 15:17:44.172599 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:17:45 crc kubenswrapper[4945]: I1008 15:17:45.938164 4945 scope.go:117] "RemoveContainer" containerID="2cf5d9b998e547999189f9001dcc7ed904fe4d92f8022806cfcb78bbbb67f302" Oct 08 15:17:46 crc kubenswrapper[4945]: I1008 15:17:46.520773 4945 scope.go:117] "RemoveContainer" containerID="61216899b7ae7ad5efd3a055febd62e6c8108b6059bf69434cf1930180061080" Oct 08 15:17:46 crc kubenswrapper[4945]: I1008 15:17:46.768032 4945 scope.go:117] "RemoveContainer" containerID="5fb9b94daeaa462fd3c4a5e750576d711a4f878418242bffe0aa1bdb55ab4b3e" Oct 08 15:17:46 crc kubenswrapper[4945]: I1008 15:17:46.889730 4945 scope.go:117] "RemoveContainer" containerID="520954aa2a83fb1a9bcb6a80baf3e60d552de350615a010c9ab8653fb42f178b" Oct 08 15:17:46 crc kubenswrapper[4945]: I1008 15:17:46.916883 4945 scope.go:117] "RemoveContainer" containerID="423ae40a4061cd10dcae509ad6219121dc5e1b1c509b7a8d5da574f277736729" Oct 08 15:17:46 crc kubenswrapper[4945]: I1008 15:17:46.951440 4945 scope.go:117] "RemoveContainer" containerID="91627156ec90af11146b7ac4df857042389fe06650fb089e2b2c12a8b15fee72" Oct 08 15:17:46 crc kubenswrapper[4945]: I1008 15:17:46.967534 4945 scope.go:117] "RemoveContainer" containerID="37e05096d57b26f4d3a94d8d0aadbd95dc220cd936b65a231a82aba79cb99b79" Oct 08 15:17:46 crc kubenswrapper[4945]: W1008 15:17:46.999854 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-d4808c616e36e55d5d4e55a26bbec9b26ad4501b26465378563c2591a1f0a4da WatchSource:0}: Error finding container d4808c616e36e55d5d4e55a26bbec9b26ad4501b26465378563c2591a1f0a4da: Status 404 returned error can't find the container with id d4808c616e36e55d5d4e55a26bbec9b26ad4501b26465378563c2591a1f0a4da Oct 08 15:17:47 crc kubenswrapper[4945]: I1008 15:17:47.207516 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d4808c616e36e55d5d4e55a26bbec9b26ad4501b26465378563c2591a1f0a4da"} Oct 08 15:17:47 crc kubenswrapper[4945]: I1008 15:17:47.209317 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fbbr4" event={"ID":"7cff81a6-1ecf-4615-a620-805f9fa659b0","Type":"ContainerStarted","Data":"9244e9ba740a95622f158814a185d98329b7082134788fbab41b440c1ada9bec"} Oct 08 15:17:47 crc kubenswrapper[4945]: I1008 15:17:47.211344 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrx2t" event={"ID":"5cfb91ef-187d-4876-bf59-d86a23eb0b74","Type":"ContainerStarted","Data":"6a66b6e25ba18d6dbd13bc4f1b2026a695622989390fe4684de4e68b77040020"} Oct 08 15:17:47 crc kubenswrapper[4945]: I1008 15:17:47.215059 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpvxx" event={"ID":"05326826-d64d-44d4-8cf0-1bd591c4fbf2","Type":"ContainerStarted","Data":"f414de872bc8c316af9514bf243ed36c06aaca21eea61045b95165cd9a578f56"} Oct 08 15:17:47 crc kubenswrapper[4945]: I1008 15:17:47.226665 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerStarted","Data":"e1b009ce30c9519794394b4a77964dafcb425a6d51c5fef8b958d9c558df5d8b"} Oct 08 15:17:47 crc kubenswrapper[4945]: W1008 15:17:47.269493 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-e2d50528614ddc6a96d58e174678f67c44df8b53c1300764d984138630d9f8be WatchSource:0}: Error finding container e2d50528614ddc6a96d58e174678f67c44df8b53c1300764d984138630d9f8be: Status 404 returned error can't find the container with id e2d50528614ddc6a96d58e174678f67c44df8b53c1300764d984138630d9f8be Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.235020 4945 generic.go:334] "Generic (PLEG): container finished" podID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerID="6a66b6e25ba18d6dbd13bc4f1b2026a695622989390fe4684de4e68b77040020" exitCode=0 Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.235154 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrx2t" event={"ID":"5cfb91ef-187d-4876-bf59-d86a23eb0b74","Type":"ContainerDied","Data":"6a66b6e25ba18d6dbd13bc4f1b2026a695622989390fe4684de4e68b77040020"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.242307 4945 generic.go:334] "Generic (PLEG): container finished" podID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerID="f414de872bc8c316af9514bf243ed36c06aaca21eea61045b95165cd9a578f56" exitCode=0 Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.242419 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpvxx" event={"ID":"05326826-d64d-44d4-8cf0-1bd591c4fbf2","Type":"ContainerDied","Data":"f414de872bc8c316af9514bf243ed36c06aaca21eea61045b95165cd9a578f56"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.245073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7cb0f8f01af7db82be7a2eb7939b77b99f8fcd5c46e8f69a297938f187724d48"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.245122 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"85a9161a8802719fa8fc3ed037d633cce4c7bc63e5f3e1d2b9f8777959e4316c"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.250570 4945 generic.go:334] "Generic (PLEG): container finished" podID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerID="e1b009ce30c9519794394b4a77964dafcb425a6d51c5fef8b958d9c558df5d8b" exitCode=0 Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.250682 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerDied","Data":"e1b009ce30c9519794394b4a77964dafcb425a6d51c5fef8b958d9c558df5d8b"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.254237 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1d0685405703eb57cb9f34240f5393af01856ea4ff78a5024dec3ecdf01429b5"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.260327 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e8422e90d0193aa8a032f72f541fd0a2d5b112e57ba211663f59bb0304c972ef"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.260411 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e2d50528614ddc6a96d58e174678f67c44df8b53c1300764d984138630d9f8be"} Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.262099 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.285019 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fbbr4" podStartSLOduration=5.068728979 podStartE2EDuration="1m50.28500059s" podCreationTimestamp="2025-10-08 15:15:58 +0000 UTC" firstStartedPulling="2025-10-08 15:16:01.480512404 +0000 UTC m=+170.834427305" lastFinishedPulling="2025-10-08 15:17:46.696783975 +0000 UTC m=+276.050698916" observedRunningTime="2025-10-08 15:17:48.281646654 +0000 UTC m=+277.635561555" watchObservedRunningTime="2025-10-08 15:17:48.28500059 +0000 UTC m=+277.638915491" Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.923120 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:17:48 crc kubenswrapper[4945]: I1008 15:17:48.923727 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:17:49 crc kubenswrapper[4945]: I1008 15:17:49.269469 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrx2t" event={"ID":"5cfb91ef-187d-4876-bf59-d86a23eb0b74","Type":"ContainerStarted","Data":"a2233bbbbb7656424294e68ad211ee54c65710eba68f45fa5ad87907c685ce5e"} Oct 08 15:17:49 crc kubenswrapper[4945]: I1008 15:17:49.301540 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rrx2t" podStartSLOduration=3.708190969 podStartE2EDuration="1m54.301512305s" podCreationTimestamp="2025-10-08 15:15:55 +0000 UTC" firstStartedPulling="2025-10-08 15:15:58.442201512 +0000 UTC m=+167.796116413" lastFinishedPulling="2025-10-08 15:17:49.035522858 +0000 UTC m=+278.389437749" observedRunningTime="2025-10-08 15:17:49.300696443 +0000 UTC m=+278.654611384" watchObservedRunningTime="2025-10-08 15:17:49.301512305 +0000 UTC m=+278.655427246" Oct 08 15:17:49 crc kubenswrapper[4945]: I1008 15:17:49.969828 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fbbr4" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="registry-server" probeResult="failure" output=< Oct 08 15:17:49 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 15:17:49 crc kubenswrapper[4945]: > Oct 08 15:17:50 crc kubenswrapper[4945]: I1008 15:17:50.277699 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpvxx" event={"ID":"05326826-d64d-44d4-8cf0-1bd591c4fbf2","Type":"ContainerStarted","Data":"f97045a955b50cfc585b71376ebb688e7cd22c8e09a99f3b17aa293d216ecd55"} Oct 08 15:17:50 crc kubenswrapper[4945]: I1008 15:17:50.295932 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rpvxx" podStartSLOduration=4.238908846 podStartE2EDuration="1m55.295914039s" podCreationTimestamp="2025-10-08 15:15:55 +0000 UTC" firstStartedPulling="2025-10-08 15:15:58.422207762 +0000 UTC m=+167.776122683" lastFinishedPulling="2025-10-08 15:17:49.479212975 +0000 UTC m=+278.833127876" observedRunningTime="2025-10-08 15:17:50.295795756 +0000 UTC m=+279.649710667" watchObservedRunningTime="2025-10-08 15:17:50.295914039 +0000 UTC m=+279.649828930" Oct 08 15:17:51 crc kubenswrapper[4945]: I1008 15:17:51.285921 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerStarted","Data":"f220a7588d00234df9c3b111b21e5517ba6d5d94ac7923f7c106ea55aafa896b"} Oct 08 15:17:51 crc kubenswrapper[4945]: I1008 15:17:51.311073 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xrqmk" podStartSLOduration=4.678245613 podStartE2EDuration="1m54.311052649s" podCreationTimestamp="2025-10-08 15:15:57 +0000 UTC" firstStartedPulling="2025-10-08 15:16:00.461327703 +0000 UTC m=+169.815242614" lastFinishedPulling="2025-10-08 15:17:50.094134749 +0000 UTC m=+279.448049650" observedRunningTime="2025-10-08 15:17:51.307958741 +0000 UTC m=+280.661873652" watchObservedRunningTime="2025-10-08 15:17:51.311052649 +0000 UTC m=+280.664967550" Oct 08 15:17:56 crc kubenswrapper[4945]: I1008 15:17:56.916739 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:17:56 crc kubenswrapper[4945]: I1008 15:17:56.918094 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:17:56 crc kubenswrapper[4945]: I1008 15:17:56.959426 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.168582 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.168963 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.217198 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.387454 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.389585 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.952709 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.952774 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:17:57 crc kubenswrapper[4945]: I1008 15:17:57.991393 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:17:58 crc kubenswrapper[4945]: I1008 15:17:58.374426 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:17:58 crc kubenswrapper[4945]: I1008 15:17:58.956390 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:17:58 crc kubenswrapper[4945]: I1008 15:17:58.990836 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:17:59 crc kubenswrapper[4945]: I1008 15:17:59.479589 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpvxx"] Oct 08 15:18:00 crc kubenswrapper[4945]: I1008 15:18:00.349701 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rpvxx" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="registry-server" containerID="cri-o://f97045a955b50cfc585b71376ebb688e7cd22c8e09a99f3b17aa293d216ecd55" gracePeriod=2 Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.383645 4945 generic.go:334] "Generic (PLEG): container finished" podID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerID="f97045a955b50cfc585b71376ebb688e7cd22c8e09a99f3b17aa293d216ecd55" exitCode=0 Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.383830 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpvxx" event={"ID":"05326826-d64d-44d4-8cf0-1bd591c4fbf2","Type":"ContainerDied","Data":"f97045a955b50cfc585b71376ebb688e7cd22c8e09a99f3b17aa293d216ecd55"} Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.452977 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.542412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-catalog-content\") pod \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.542464 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-utilities\") pod \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.542553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw7z6\" (UniqueName: \"kubernetes.io/projected/05326826-d64d-44d4-8cf0-1bd591c4fbf2-kube-api-access-qw7z6\") pod \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\" (UID: \"05326826-d64d-44d4-8cf0-1bd591c4fbf2\") " Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.543752 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-utilities" (OuterVolumeSpecName: "utilities") pod "05326826-d64d-44d4-8cf0-1bd591c4fbf2" (UID: "05326826-d64d-44d4-8cf0-1bd591c4fbf2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.564469 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05326826-d64d-44d4-8cf0-1bd591c4fbf2-kube-api-access-qw7z6" (OuterVolumeSpecName: "kube-api-access-qw7z6") pod "05326826-d64d-44d4-8cf0-1bd591c4fbf2" (UID: "05326826-d64d-44d4-8cf0-1bd591c4fbf2"). InnerVolumeSpecName "kube-api-access-qw7z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.600313 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05326826-d64d-44d4-8cf0-1bd591c4fbf2" (UID: "05326826-d64d-44d4-8cf0-1bd591c4fbf2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.644103 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw7z6\" (UniqueName: \"kubernetes.io/projected/05326826-d64d-44d4-8cf0-1bd591c4fbf2-kube-api-access-qw7z6\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.644152 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:04 crc kubenswrapper[4945]: I1008 15:18:04.644160 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05326826-d64d-44d4-8cf0-1bd591c4fbf2-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:05 crc kubenswrapper[4945]: I1008 15:18:05.390898 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpvxx" event={"ID":"05326826-d64d-44d4-8cf0-1bd591c4fbf2","Type":"ContainerDied","Data":"c51f37a84a5865d9a71ceeadc7834ea347630069f45486f981e83b8d3b8b0d86"} Oct 08 15:18:05 crc kubenswrapper[4945]: I1008 15:18:05.390959 4945 scope.go:117] "RemoveContainer" containerID="f97045a955b50cfc585b71376ebb688e7cd22c8e09a99f3b17aa293d216ecd55" Oct 08 15:18:05 crc kubenswrapper[4945]: I1008 15:18:05.390991 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpvxx" Oct 08 15:18:05 crc kubenswrapper[4945]: I1008 15:18:05.418850 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpvxx"] Oct 08 15:18:05 crc kubenswrapper[4945]: I1008 15:18:05.423846 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rpvxx"] Oct 08 15:18:05 crc kubenswrapper[4945]: I1008 15:18:05.425925 4945 scope.go:117] "RemoveContainer" containerID="f414de872bc8c316af9514bf243ed36c06aaca21eea61045b95165cd9a578f56" Oct 08 15:18:05 crc kubenswrapper[4945]: I1008 15:18:05.442459 4945 scope.go:117] "RemoveContainer" containerID="899985b52404eb307d0963509f8e01a870abbc983349a9c778342bb538008bb6" Oct 08 15:18:06 crc kubenswrapper[4945]: I1008 15:18:06.034855 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" path="/var/lib/kubelet/pods/05326826-d64d-44d4-8cf0-1bd591c4fbf2/volumes" Oct 08 15:18:17 crc kubenswrapper[4945]: I1008 15:18:17.137257 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-d8prt"] Oct 08 15:18:24 crc kubenswrapper[4945]: I1008 15:18:24.179349 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.165375 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" podUID="8db81fa8-6d99-4529-a1a8-c573752dfb18" containerName="oauth-openshift" containerID="cri-o://05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e" gracePeriod=15 Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.531064 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581229 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-d787499bd-k96nx"] Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581519 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581536 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581546 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581556 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581572 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581582 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581597 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581606 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581619 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581628 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581640 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8db81fa8-6d99-4529-a1a8-c573752dfb18" containerName="oauth-openshift" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581648 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8db81fa8-6d99-4529-a1a8-c573752dfb18" containerName="oauth-openshift" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581658 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba793aa-5158-46a3-8ea6-1e91389524d1" containerName="collect-profiles" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581666 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba793aa-5158-46a3-8ea6-1e91389524d1" containerName="collect-profiles" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581681 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581689 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581702 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581710 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581722 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581732 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581744 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581753 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581766 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581775 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="extract-content" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581797 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581805 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581816 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581826 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="extract-utilities" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.581839 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54cb337-afd0-4ced-832e-62578051f906" containerName="pruner" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581847 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54cb337-afd0-4ced-832e-62578051f906" containerName="pruner" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581983 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54cb337-afd0-4ced-832e-62578051f906" containerName="pruner" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.581996 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcb19c22-f621-450d-b675-1b2c3a1c4dbc" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.582005 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="05326826-d64d-44d4-8cf0-1bd591c4fbf2" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.582017 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8db81fa8-6d99-4529-a1a8-c573752dfb18" containerName="oauth-openshift" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.582028 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcdc5718-c74a-4abc-80f7-d7de2bd829d1" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.582038 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ed9dfc5-a0bf-47c2-b32e-68a5bd1449c7" containerName="registry-server" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.582053 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba793aa-5158-46a3-8ea6-1e91389524d1" containerName="collect-profiles" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.582620 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.589325 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d787499bd-k96nx"] Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.606364 4945 generic.go:334] "Generic (PLEG): container finished" podID="8db81fa8-6d99-4529-a1a8-c573752dfb18" containerID="05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e" exitCode=0 Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.606425 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" event={"ID":"8db81fa8-6d99-4529-a1a8-c573752dfb18","Type":"ContainerDied","Data":"05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e"} Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.606450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" event={"ID":"8db81fa8-6d99-4529-a1a8-c573752dfb18","Type":"ContainerDied","Data":"fc3bd2630277d9e53ec5b55b53e11506ad3d22d14933b6aae1b5adab6ac3136d"} Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.606486 4945 scope.go:117] "RemoveContainer" containerID="05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.606824 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-d8prt" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.624539 4945 scope.go:117] "RemoveContainer" containerID="05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e" Oct 08 15:18:42 crc kubenswrapper[4945]: E1008 15:18:42.625011 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e\": container with ID starting with 05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e not found: ID does not exist" containerID="05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.625044 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e"} err="failed to get container status \"05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e\": rpc error: code = NotFound desc = could not find container \"05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e\": container with ID starting with 05b2f1865ad018f5d499d838b1ab8b5d5ca01a5c0570e83c3600c4a6763dc35e not found: ID does not exist" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.648801 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-trusted-ca-bundle\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.648852 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-serving-cert\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.648882 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-session\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.648951 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-dir\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.648979 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-router-certs\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649008 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-cliconfig\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649029 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-error\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649060 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc4qf\" (UniqueName: \"kubernetes.io/projected/8db81fa8-6d99-4529-a1a8-c573752dfb18-kube-api-access-sc4qf\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-service-ca\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649148 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-ocp-branding-template\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649172 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-policies\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649192 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-login\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649209 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-idp-0-file-data\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649239 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-provider-selection\") pod \"8db81fa8-6d99-4529-a1a8-c573752dfb18\" (UID: \"8db81fa8-6d99-4529-a1a8-c573752dfb18\") " Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.649648 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.650092 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.650756 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.650897 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.651334 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.655799 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.656047 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.656088 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8db81fa8-6d99-4529-a1a8-c573752dfb18-kube-api-access-sc4qf" (OuterVolumeSpecName: "kube-api-access-sc4qf") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "kube-api-access-sc4qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.656489 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.656643 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.656933 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.657637 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.661651 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.662417 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "8db81fa8-6d99-4529-a1a8-c573752dfb18" (UID: "8db81fa8-6d99-4529-a1a8-c573752dfb18"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.750815 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-login\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.750901 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.750951 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.750990 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751026 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751078 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-session\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751171 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751220 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-error\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751438 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlvl7\" (UniqueName: \"kubernetes.io/projected/5001e8f3-5567-4903-b329-803fb941a6d3-kube-api-access-tlvl7\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751597 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5001e8f3-5567-4903-b329-803fb941a6d3-audit-dir\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751673 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751750 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-audit-policies\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.751912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752051 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752075 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752091 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752105 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752139 4945 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752152 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752164 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752176 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752192 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752204 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc4qf\" (UniqueName: \"kubernetes.io/projected/8db81fa8-6d99-4529-a1a8-c573752dfb18-kube-api-access-sc4qf\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752217 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752226 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8db81fa8-6d99-4529-a1a8-c573752dfb18-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752238 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.752253 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8db81fa8-6d99-4529-a1a8-c573752dfb18-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853436 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5001e8f3-5567-4903-b329-803fb941a6d3-audit-dir\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853495 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853518 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853539 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-audit-policies\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853568 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853583 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-login\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5001e8f3-5567-4903-b329-803fb941a6d3-audit-dir\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.853600 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.854341 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.854593 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855162 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855231 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855317 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855272 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855576 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-session\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855727 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855870 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-error\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.855726 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5001e8f3-5567-4903-b329-803fb941a6d3-audit-policies\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.856082 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlvl7\" (UniqueName: \"kubernetes.io/projected/5001e8f3-5567-4903-b329-803fb941a6d3-kube-api-access-tlvl7\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.858735 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-login\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.859933 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-error\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.860028 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.860104 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.860405 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-session\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.860597 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.861267 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.861638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5001e8f3-5567-4903-b329-803fb941a6d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.876622 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlvl7\" (UniqueName: \"kubernetes.io/projected/5001e8f3-5567-4903-b329-803fb941a6d3-kube-api-access-tlvl7\") pod \"oauth-openshift-d787499bd-k96nx\" (UID: \"5001e8f3-5567-4903-b329-803fb941a6d3\") " pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.898910 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.943164 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-d8prt"] Oct 08 15:18:42 crc kubenswrapper[4945]: I1008 15:18:42.948482 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-d8prt"] Oct 08 15:18:43 crc kubenswrapper[4945]: I1008 15:18:43.100190 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d787499bd-k96nx"] Oct 08 15:18:43 crc kubenswrapper[4945]: I1008 15:18:43.614363 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" event={"ID":"5001e8f3-5567-4903-b329-803fb941a6d3","Type":"ContainerStarted","Data":"896fd6c60965103c3487f82ae44135f70f287853e605dc6e223db2bf3f7121d1"} Oct 08 15:18:43 crc kubenswrapper[4945]: I1008 15:18:43.614415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" event={"ID":"5001e8f3-5567-4903-b329-803fb941a6d3","Type":"ContainerStarted","Data":"d6935162f26fdd92bff241073b1028d020bd983be4052f4275a77258ed0f7091"} Oct 08 15:18:43 crc kubenswrapper[4945]: I1008 15:18:43.614553 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:43 crc kubenswrapper[4945]: I1008 15:18:43.637532 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" podStartSLOduration=26.637509773 podStartE2EDuration="26.637509773s" podCreationTimestamp="2025-10-08 15:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:18:43.633618968 +0000 UTC m=+332.987533869" watchObservedRunningTime="2025-10-08 15:18:43.637509773 +0000 UTC m=+332.991424674" Oct 08 15:18:43 crc kubenswrapper[4945]: I1008 15:18:43.700709 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-d787499bd-k96nx" Oct 08 15:18:44 crc kubenswrapper[4945]: I1008 15:18:44.033821 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8db81fa8-6d99-4529-a1a8-c573752dfb18" path="/var/lib/kubelet/pods/8db81fa8-6d99-4529-a1a8-c573752dfb18/volumes" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.282784 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rrx2t"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.283613 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rrx2t" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="registry-server" containerID="cri-o://a2233bbbbb7656424294e68ad211ee54c65710eba68f45fa5ad87907c685ce5e" gracePeriod=30 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.297043 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gx9hj"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.297329 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gx9hj" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="registry-server" containerID="cri-o://a593a009cca0af16b9859f8bb0c7a75a9caa8eced37d04d5d69dc502ebb28800" gracePeriod=30 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.309144 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m5fwt"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.309478 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" podUID="e76deaf4-7012-4889-87bf-1310e42a6f46" containerName="marketplace-operator" containerID="cri-o://478229d09b8b505553a9ffdf2715f3815d1a0ec868c769a7f07f319ef23f16f0" gracePeriod=30 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.320510 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrqmk"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.320779 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xrqmk" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="registry-server" containerID="cri-o://f220a7588d00234df9c3b111b21e5517ba6d5d94ac7923f7c106ea55aafa896b" gracePeriod=30 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.329618 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vnm62"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.330531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.335324 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fbbr4"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.335589 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fbbr4" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="registry-server" containerID="cri-o://9244e9ba740a95622f158814a185d98329b7082134788fbab41b440c1ada9bec" gracePeriod=30 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.345281 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vnm62"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.499830 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e99374e4-2a46-4044-881c-9bdb32f4a215-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.500025 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e99374e4-2a46-4044-881c-9bdb32f4a215-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.500053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7zlf\" (UniqueName: \"kubernetes.io/projected/e99374e4-2a46-4044-881c-9bdb32f4a215-kube-api-access-l7zlf\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.601153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e99374e4-2a46-4044-881c-9bdb32f4a215-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.601476 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7zlf\" (UniqueName: \"kubernetes.io/projected/e99374e4-2a46-4044-881c-9bdb32f4a215-kube-api-access-l7zlf\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.601517 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e99374e4-2a46-4044-881c-9bdb32f4a215-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.603030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e99374e4-2a46-4044-881c-9bdb32f4a215-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.609210 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e99374e4-2a46-4044-881c-9bdb32f4a215-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.617007 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7zlf\" (UniqueName: \"kubernetes.io/projected/e99374e4-2a46-4044-881c-9bdb32f4a215-kube-api-access-l7zlf\") pod \"marketplace-operator-79b997595-vnm62\" (UID: \"e99374e4-2a46-4044-881c-9bdb32f4a215\") " pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.651515 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.689168 4945 generic.go:334] "Generic (PLEG): container finished" podID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerID="a2233bbbbb7656424294e68ad211ee54c65710eba68f45fa5ad87907c685ce5e" exitCode=0 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.689229 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrx2t" event={"ID":"5cfb91ef-187d-4876-bf59-d86a23eb0b74","Type":"ContainerDied","Data":"a2233bbbbb7656424294e68ad211ee54c65710eba68f45fa5ad87907c685ce5e"} Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.689255 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrx2t" event={"ID":"5cfb91ef-187d-4876-bf59-d86a23eb0b74","Type":"ContainerDied","Data":"53ea517aaab9c1daf253a4c6b62023a569fcfab7faf3f58908b2476ab2982386"} Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.689266 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53ea517aaab9c1daf253a4c6b62023a569fcfab7faf3f58908b2476ab2982386" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.692191 4945 generic.go:334] "Generic (PLEG): container finished" podID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerID="f220a7588d00234df9c3b111b21e5517ba6d5d94ac7923f7c106ea55aafa896b" exitCode=0 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.692256 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerDied","Data":"f220a7588d00234df9c3b111b21e5517ba6d5d94ac7923f7c106ea55aafa896b"} Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.699682 4945 generic.go:334] "Generic (PLEG): container finished" podID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerID="9244e9ba740a95622f158814a185d98329b7082134788fbab41b440c1ada9bec" exitCode=0 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.699775 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fbbr4" event={"ID":"7cff81a6-1ecf-4615-a620-805f9fa659b0","Type":"ContainerDied","Data":"9244e9ba740a95622f158814a185d98329b7082134788fbab41b440c1ada9bec"} Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.703602 4945 generic.go:334] "Generic (PLEG): container finished" podID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerID="a593a009cca0af16b9859f8bb0c7a75a9caa8eced37d04d5d69dc502ebb28800" exitCode=0 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.703674 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx9hj" event={"ID":"0ac7a743-3ca9-4a2f-832b-3392d945b7b5","Type":"ContainerDied","Data":"a593a009cca0af16b9859f8bb0c7a75a9caa8eced37d04d5d69dc502ebb28800"} Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.705234 4945 generic.go:334] "Generic (PLEG): container finished" podID="e76deaf4-7012-4889-87bf-1310e42a6f46" containerID="478229d09b8b505553a9ffdf2715f3815d1a0ec868c769a7f07f319ef23f16f0" exitCode=0 Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.705261 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" event={"ID":"e76deaf4-7012-4889-87bf-1310e42a6f46","Type":"ContainerDied","Data":"478229d09b8b505553a9ffdf2715f3815d1a0ec868c769a7f07f319ef23f16f0"} Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.744553 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.751218 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.765178 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.787290 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.790856 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.907206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sww7c\" (UniqueName: \"kubernetes.io/projected/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-kube-api-access-sww7c\") pod \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.907325 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-utilities\") pod \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.908592 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-utilities" (OuterVolumeSpecName: "utilities") pod "2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" (UID: "2cf6d252-89f0-46bc-a4aa-5c2867e3baf3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.908689 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcnck\" (UniqueName: \"kubernetes.io/projected/5cfb91ef-187d-4876-bf59-d86a23eb0b74-kube-api-access-qcnck\") pod \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909190 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-catalog-content\") pod \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909265 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-trusted-ca\") pod \"e76deaf4-7012-4889-87bf-1310e42a6f46\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909299 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-catalog-content\") pod \"7cff81a6-1ecf-4615-a620-805f9fa659b0\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909338 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-utilities\") pod \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\" (UID: \"0ac7a743-3ca9-4a2f-832b-3392d945b7b5\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-utilities\") pod \"7cff81a6-1ecf-4615-a620-805f9fa659b0\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzjs8\" (UniqueName: \"kubernetes.io/projected/7cff81a6-1ecf-4615-a620-805f9fa659b0-kube-api-access-nzjs8\") pod \"7cff81a6-1ecf-4615-a620-805f9fa659b0\" (UID: \"7cff81a6-1ecf-4615-a620-805f9fa659b0\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909452 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw27x\" (UniqueName: \"kubernetes.io/projected/e76deaf4-7012-4889-87bf-1310e42a6f46-kube-api-access-tw27x\") pod \"e76deaf4-7012-4889-87bf-1310e42a6f46\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909504 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55hsq\" (UniqueName: \"kubernetes.io/projected/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-kube-api-access-55hsq\") pod \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909558 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-utilities\") pod \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909758 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-catalog-content\") pod \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\" (UID: \"5cfb91ef-187d-4876-bf59-d86a23eb0b74\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909797 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-operator-metrics\") pod \"e76deaf4-7012-4889-87bf-1310e42a6f46\" (UID: \"e76deaf4-7012-4889-87bf-1310e42a6f46\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.909831 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-catalog-content\") pod \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\" (UID: \"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3\") " Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.910215 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.911570 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-utilities" (OuterVolumeSpecName: "utilities") pod "5cfb91ef-187d-4876-bf59-d86a23eb0b74" (UID: "5cfb91ef-187d-4876-bf59-d86a23eb0b74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.912346 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-utilities" (OuterVolumeSpecName: "utilities") pod "0ac7a743-3ca9-4a2f-832b-3392d945b7b5" (UID: "0ac7a743-3ca9-4a2f-832b-3392d945b7b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.912413 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "e76deaf4-7012-4889-87bf-1310e42a6f46" (UID: "e76deaf4-7012-4889-87bf-1310e42a6f46"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.912871 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-utilities" (OuterVolumeSpecName: "utilities") pod "7cff81a6-1ecf-4615-a620-805f9fa659b0" (UID: "7cff81a6-1ecf-4615-a620-805f9fa659b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.913225 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cfb91ef-187d-4876-bf59-d86a23eb0b74-kube-api-access-qcnck" (OuterVolumeSpecName: "kube-api-access-qcnck") pod "5cfb91ef-187d-4876-bf59-d86a23eb0b74" (UID: "5cfb91ef-187d-4876-bf59-d86a23eb0b74"). InnerVolumeSpecName "kube-api-access-qcnck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.913465 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cff81a6-1ecf-4615-a620-805f9fa659b0-kube-api-access-nzjs8" (OuterVolumeSpecName: "kube-api-access-nzjs8") pod "7cff81a6-1ecf-4615-a620-805f9fa659b0" (UID: "7cff81a6-1ecf-4615-a620-805f9fa659b0"). InnerVolumeSpecName "kube-api-access-nzjs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.914433 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-kube-api-access-sww7c" (OuterVolumeSpecName: "kube-api-access-sww7c") pod "0ac7a743-3ca9-4a2f-832b-3392d945b7b5" (UID: "0ac7a743-3ca9-4a2f-832b-3392d945b7b5"). InnerVolumeSpecName "kube-api-access-sww7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.916914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-kube-api-access-55hsq" (OuterVolumeSpecName: "kube-api-access-55hsq") pod "2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" (UID: "2cf6d252-89f0-46bc-a4aa-5c2867e3baf3"). InnerVolumeSpecName "kube-api-access-55hsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.918167 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "e76deaf4-7012-4889-87bf-1310e42a6f46" (UID: "e76deaf4-7012-4889-87bf-1310e42a6f46"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.921767 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e76deaf4-7012-4889-87bf-1310e42a6f46-kube-api-access-tw27x" (OuterVolumeSpecName: "kube-api-access-tw27x") pod "e76deaf4-7012-4889-87bf-1310e42a6f46" (UID: "e76deaf4-7012-4889-87bf-1310e42a6f46"). InnerVolumeSpecName "kube-api-access-tw27x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.935066 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" (UID: "2cf6d252-89f0-46bc-a4aa-5c2867e3baf3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.951175 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vnm62"] Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.990755 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cfb91ef-187d-4876-bf59-d86a23eb0b74" (UID: "5cfb91ef-187d-4876-bf59-d86a23eb0b74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:54 crc kubenswrapper[4945]: I1008 15:18:54.992584 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ac7a743-3ca9-4a2f-832b-3392d945b7b5" (UID: "0ac7a743-3ca9-4a2f-832b-3392d945b7b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.002026 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7cff81a6-1ecf-4615-a620-805f9fa659b0" (UID: "7cff81a6-1ecf-4615-a620-805f9fa659b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011666 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011698 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011712 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011725 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011736 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cff81a6-1ecf-4615-a620-805f9fa659b0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011747 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzjs8\" (UniqueName: \"kubernetes.io/projected/7cff81a6-1ecf-4615-a620-805f9fa659b0-kube-api-access-nzjs8\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011759 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw27x\" (UniqueName: \"kubernetes.io/projected/e76deaf4-7012-4889-87bf-1310e42a6f46-kube-api-access-tw27x\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011770 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55hsq\" (UniqueName: \"kubernetes.io/projected/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-kube-api-access-55hsq\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011782 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011792 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cfb91ef-187d-4876-bf59-d86a23eb0b74-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011803 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e76deaf4-7012-4889-87bf-1310e42a6f46-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011817 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011830 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sww7c\" (UniqueName: \"kubernetes.io/projected/0ac7a743-3ca9-4a2f-832b-3392d945b7b5-kube-api-access-sww7c\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.011842 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcnck\" (UniqueName: \"kubernetes.io/projected/5cfb91ef-187d-4876-bf59-d86a23eb0b74-kube-api-access-qcnck\") on node \"crc\" DevicePath \"\"" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.712375 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fbbr4" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.712349 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fbbr4" event={"ID":"7cff81a6-1ecf-4615-a620-805f9fa659b0","Type":"ContainerDied","Data":"b552e010444dd001dd3662c45c81fa4445946afdd90fc51b398cc6bec65c166e"} Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.712523 4945 scope.go:117] "RemoveContainer" containerID="9244e9ba740a95622f158814a185d98329b7082134788fbab41b440c1ada9bec" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.715556 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx9hj" event={"ID":"0ac7a743-3ca9-4a2f-832b-3392d945b7b5","Type":"ContainerDied","Data":"40fb20db8ff7f5ef418098184bc6eb89d57f5c14c61d3175b02c3762f2a0151f"} Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.715638 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx9hj" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.728676 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.728788 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m5fwt" event={"ID":"e76deaf4-7012-4889-87bf-1310e42a6f46","Type":"ContainerDied","Data":"cac52278a6d6e278a45b93bc8f09be36005d6d227bd18251bd86b6fac185a8b5"} Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.730821 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" event={"ID":"e99374e4-2a46-4044-881c-9bdb32f4a215","Type":"ContainerStarted","Data":"9247cdd9b05a451b1e13dacd895dcb9690ce93384da471a9300f46c5ecb879fb"} Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.730973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" event={"ID":"e99374e4-2a46-4044-881c-9bdb32f4a215","Type":"ContainerStarted","Data":"417bb5c6d03bc701ce961b1194f4bcf0c62f7c730976bc8e867603de258d10d6"} Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.731565 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.740438 4945 scope.go:117] "RemoveContainer" containerID="35b7762f9fd5cd43660f4ad32a3fa17327a43685177d828a2cdddb754f658a2a" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.740541 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.741699 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrx2t" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.741711 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrqmk" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.741664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrqmk" event={"ID":"2cf6d252-89f0-46bc-a4aa-5c2867e3baf3","Type":"ContainerDied","Data":"aa3ffd9978ca54d7b316834ca782635ac0820ebac5287a8b763839ee486e6184"} Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.763389 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fbbr4"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.771918 4945 scope.go:117] "RemoveContainer" containerID="95a6a9f989163672933e0b2d62e1e821271c1aa464d8ae723e0bfdc9b9e0d935" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.772330 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fbbr4"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.821467 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vnm62" podStartSLOduration=1.821444332 podStartE2EDuration="1.821444332s" podCreationTimestamp="2025-10-08 15:18:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:18:55.792805201 +0000 UTC m=+345.146720102" watchObservedRunningTime="2025-10-08 15:18:55.821444332 +0000 UTC m=+345.175359243" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.825744 4945 scope.go:117] "RemoveContainer" containerID="a593a009cca0af16b9859f8bb0c7a75a9caa8eced37d04d5d69dc502ebb28800" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.827281 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrqmk"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.837276 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrqmk"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.843753 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m5fwt"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.844003 4945 scope.go:117] "RemoveContainer" containerID="11fa19a09828a5d4272fa5bb0f742e5782770a63eddc8cf146aef739df92fc82" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.850414 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m5fwt"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.855168 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gx9hj"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.861462 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gx9hj"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.866243 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rrx2t"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.874485 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rrx2t"] Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.877295 4945 scope.go:117] "RemoveContainer" containerID="8c6e0224ce05d4717d3e81960458f58378dbb7774acb671ce1f7d37d69e941d1" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.906717 4945 scope.go:117] "RemoveContainer" containerID="478229d09b8b505553a9ffdf2715f3815d1a0ec868c769a7f07f319ef23f16f0" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.934009 4945 scope.go:117] "RemoveContainer" containerID="f220a7588d00234df9c3b111b21e5517ba6d5d94ac7923f7c106ea55aafa896b" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.952040 4945 scope.go:117] "RemoveContainer" containerID="e1b009ce30c9519794394b4a77964dafcb425a6d51c5fef8b958d9c558df5d8b" Oct 08 15:18:55 crc kubenswrapper[4945]: I1008 15:18:55.966496 4945 scope.go:117] "RemoveContainer" containerID="d5a2e88ad436714c340076f875b1c32c18b76f6a5fdddcf2d201a61e34857675" Oct 08 15:18:56 crc kubenswrapper[4945]: I1008 15:18:56.031596 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" path="/var/lib/kubelet/pods/0ac7a743-3ca9-4a2f-832b-3392d945b7b5/volumes" Oct 08 15:18:56 crc kubenswrapper[4945]: I1008 15:18:56.032863 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" path="/var/lib/kubelet/pods/2cf6d252-89f0-46bc-a4aa-5c2867e3baf3/volumes" Oct 08 15:18:56 crc kubenswrapper[4945]: I1008 15:18:56.034037 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" path="/var/lib/kubelet/pods/5cfb91ef-187d-4876-bf59-d86a23eb0b74/volumes" Oct 08 15:18:56 crc kubenswrapper[4945]: I1008 15:18:56.036036 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" path="/var/lib/kubelet/pods/7cff81a6-1ecf-4615-a620-805f9fa659b0/volumes" Oct 08 15:18:56 crc kubenswrapper[4945]: I1008 15:18:56.037298 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76deaf4-7012-4889-87bf-1310e42a6f46" path="/var/lib/kubelet/pods/e76deaf4-7012-4889-87bf-1310e42a6f46/volumes" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.089403 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9xdxb"] Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090063 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090122 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090136 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090143 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090155 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090164 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090176 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090183 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090194 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090201 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090212 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090220 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090230 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76deaf4-7012-4889-87bf-1310e42a6f46" containerName="marketplace-operator" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090237 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76deaf4-7012-4889-87bf-1310e42a6f46" containerName="marketplace-operator" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090246 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090252 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090259 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090265 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090270 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090306 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="extract-content" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090315 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090320 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090328 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090333 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="extract-utilities" Oct 08 15:18:57 crc kubenswrapper[4945]: E1008 15:18:57.090343 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090348 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090523 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cff81a6-1ecf-4615-a620-805f9fa659b0" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090537 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf6d252-89f0-46bc-a4aa-5c2867e3baf3" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090545 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76deaf4-7012-4889-87bf-1310e42a6f46" containerName="marketplace-operator" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090554 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac7a743-3ca9-4a2f-832b-3392d945b7b5" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.090561 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cfb91ef-187d-4876-bf59-d86a23eb0b74" containerName="registry-server" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.096174 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.097637 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9xdxb"] Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.099218 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.241578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppnzr\" (UniqueName: \"kubernetes.io/projected/db0ecadd-b135-46e5-aa0e-92d952913a11-kube-api-access-ppnzr\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.241689 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0ecadd-b135-46e5-aa0e-92d952913a11-utilities\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.241943 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0ecadd-b135-46e5-aa0e-92d952913a11-catalog-content\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.342869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0ecadd-b135-46e5-aa0e-92d952913a11-catalog-content\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.342938 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppnzr\" (UniqueName: \"kubernetes.io/projected/db0ecadd-b135-46e5-aa0e-92d952913a11-kube-api-access-ppnzr\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.342961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0ecadd-b135-46e5-aa0e-92d952913a11-utilities\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.343544 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0ecadd-b135-46e5-aa0e-92d952913a11-utilities\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.343820 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0ecadd-b135-46e5-aa0e-92d952913a11-catalog-content\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.376084 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppnzr\" (UniqueName: \"kubernetes.io/projected/db0ecadd-b135-46e5-aa0e-92d952913a11-kube-api-access-ppnzr\") pod \"redhat-operators-9xdxb\" (UID: \"db0ecadd-b135-46e5-aa0e-92d952913a11\") " pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.412972 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.596771 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9xdxb"] Oct 08 15:18:57 crc kubenswrapper[4945]: W1008 15:18:57.602631 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb0ecadd_b135_46e5_aa0e_92d952913a11.slice/crio-fd876d772595dd003c9870e64acf42760d2cd9554867ce9810cb30db60cd5bc5 WatchSource:0}: Error finding container fd876d772595dd003c9870e64acf42760d2cd9554867ce9810cb30db60cd5bc5: Status 404 returned error can't find the container with id fd876d772595dd003c9870e64acf42760d2cd9554867ce9810cb30db60cd5bc5 Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.755466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xdxb" event={"ID":"db0ecadd-b135-46e5-aa0e-92d952913a11","Type":"ContainerStarted","Data":"104acf9c425e52746643aa039986afc5c2545550c2a54a2579ea6ecb3c732b00"} Oct 08 15:18:57 crc kubenswrapper[4945]: I1008 15:18:57.755510 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xdxb" event={"ID":"db0ecadd-b135-46e5-aa0e-92d952913a11","Type":"ContainerStarted","Data":"fd876d772595dd003c9870e64acf42760d2cd9554867ce9810cb30db60cd5bc5"} Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.091508 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rzsp2"] Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.092440 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.094614 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.112049 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rzsp2"] Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.254245 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54859f-2831-49c8-88db-4b0c4451f3b0-utilities\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.254585 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z2mb\" (UniqueName: \"kubernetes.io/projected/9c54859f-2831-49c8-88db-4b0c4451f3b0-kube-api-access-5z2mb\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.254606 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54859f-2831-49c8-88db-4b0c4451f3b0-catalog-content\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.356021 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z2mb\" (UniqueName: \"kubernetes.io/projected/9c54859f-2831-49c8-88db-4b0c4451f3b0-kube-api-access-5z2mb\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.356089 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54859f-2831-49c8-88db-4b0c4451f3b0-catalog-content\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.356188 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54859f-2831-49c8-88db-4b0c4451f3b0-utilities\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.356590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54859f-2831-49c8-88db-4b0c4451f3b0-catalog-content\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.356788 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54859f-2831-49c8-88db-4b0c4451f3b0-utilities\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.385921 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z2mb\" (UniqueName: \"kubernetes.io/projected/9c54859f-2831-49c8-88db-4b0c4451f3b0-kube-api-access-5z2mb\") pod \"community-operators-rzsp2\" (UID: \"9c54859f-2831-49c8-88db-4b0c4451f3b0\") " pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.426665 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.630778 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rzsp2"] Oct 08 15:18:58 crc kubenswrapper[4945]: W1008 15:18:58.637460 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c54859f_2831_49c8_88db_4b0c4451f3b0.slice/crio-66813c665eba355475f6167724ae96ca2f855e7e2e3e71326c16000985b1b77d WatchSource:0}: Error finding container 66813c665eba355475f6167724ae96ca2f855e7e2e3e71326c16000985b1b77d: Status 404 returned error can't find the container with id 66813c665eba355475f6167724ae96ca2f855e7e2e3e71326c16000985b1b77d Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.763581 4945 generic.go:334] "Generic (PLEG): container finished" podID="db0ecadd-b135-46e5-aa0e-92d952913a11" containerID="104acf9c425e52746643aa039986afc5c2545550c2a54a2579ea6ecb3c732b00" exitCode=0 Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.763886 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xdxb" event={"ID":"db0ecadd-b135-46e5-aa0e-92d952913a11","Type":"ContainerDied","Data":"104acf9c425e52746643aa039986afc5c2545550c2a54a2579ea6ecb3c732b00"} Oct 08 15:18:58 crc kubenswrapper[4945]: I1008 15:18:58.766776 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rzsp2" event={"ID":"9c54859f-2831-49c8-88db-4b0c4451f3b0","Type":"ContainerStarted","Data":"66813c665eba355475f6167724ae96ca2f855e7e2e3e71326c16000985b1b77d"} Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.499256 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cxqrl"] Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.501459 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.503551 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cxqrl"] Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.504391 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.671757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7bde639-beb8-4042-9adc-ad1dcde01cfc-catalog-content\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.671869 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7bde639-beb8-4042-9adc-ad1dcde01cfc-utilities\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.671896 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zch59\" (UniqueName: \"kubernetes.io/projected/a7bde639-beb8-4042-9adc-ad1dcde01cfc-kube-api-access-zch59\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.771830 4945 generic.go:334] "Generic (PLEG): container finished" podID="9c54859f-2831-49c8-88db-4b0c4451f3b0" containerID="d949503dca4d20507d4df35b2886fb8e96a4a7500ab3c543e7c145a5e7175e56" exitCode=0 Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.771883 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rzsp2" event={"ID":"9c54859f-2831-49c8-88db-4b0c4451f3b0","Type":"ContainerDied","Data":"d949503dca4d20507d4df35b2886fb8e96a4a7500ab3c543e7c145a5e7175e56"} Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.772559 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7bde639-beb8-4042-9adc-ad1dcde01cfc-catalog-content\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.772652 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zch59\" (UniqueName: \"kubernetes.io/projected/a7bde639-beb8-4042-9adc-ad1dcde01cfc-kube-api-access-zch59\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.772681 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7bde639-beb8-4042-9adc-ad1dcde01cfc-utilities\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.773034 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7bde639-beb8-4042-9adc-ad1dcde01cfc-catalog-content\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.773194 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7bde639-beb8-4042-9adc-ad1dcde01cfc-utilities\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.793142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zch59\" (UniqueName: \"kubernetes.io/projected/a7bde639-beb8-4042-9adc-ad1dcde01cfc-kube-api-access-zch59\") pod \"certified-operators-cxqrl\" (UID: \"a7bde639-beb8-4042-9adc-ad1dcde01cfc\") " pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.820434 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:18:59 crc kubenswrapper[4945]: I1008 15:18:59.989394 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cxqrl"] Oct 08 15:18:59 crc kubenswrapper[4945]: W1008 15:18:59.997197 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7bde639_beb8_4042_9adc_ad1dcde01cfc.slice/crio-54ac2779ffacc777c08b23cc72d4b0f927f58562eaf1f02b36be3f794a730333 WatchSource:0}: Error finding container 54ac2779ffacc777c08b23cc72d4b0f927f58562eaf1f02b36be3f794a730333: Status 404 returned error can't find the container with id 54ac2779ffacc777c08b23cc72d4b0f927f58562eaf1f02b36be3f794a730333 Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.493706 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-64lj5"] Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.495048 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.497940 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.502502 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-64lj5"] Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.582929 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff2444c-a4f4-4779-a11c-968c0ed65815-catalog-content\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.582994 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77gt2\" (UniqueName: \"kubernetes.io/projected/6ff2444c-a4f4-4779-a11c-968c0ed65815-kube-api-access-77gt2\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.583032 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff2444c-a4f4-4779-a11c-968c0ed65815-utilities\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.683913 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff2444c-a4f4-4779-a11c-968c0ed65815-utilities\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.683974 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff2444c-a4f4-4779-a11c-968c0ed65815-catalog-content\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.684018 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77gt2\" (UniqueName: \"kubernetes.io/projected/6ff2444c-a4f4-4779-a11c-968c0ed65815-kube-api-access-77gt2\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.684438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff2444c-a4f4-4779-a11c-968c0ed65815-utilities\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.684569 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff2444c-a4f4-4779-a11c-968c0ed65815-catalog-content\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.708628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77gt2\" (UniqueName: \"kubernetes.io/projected/6ff2444c-a4f4-4779-a11c-968c0ed65815-kube-api-access-77gt2\") pod \"redhat-marketplace-64lj5\" (UID: \"6ff2444c-a4f4-4779-a11c-968c0ed65815\") " pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.784370 4945 generic.go:334] "Generic (PLEG): container finished" podID="db0ecadd-b135-46e5-aa0e-92d952913a11" containerID="9df53b82937b2220bfa13ca28d5eec95e9e24a4b3de74a89ad08d925088eacb8" exitCode=0 Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.784727 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xdxb" event={"ID":"db0ecadd-b135-46e5-aa0e-92d952913a11","Type":"ContainerDied","Data":"9df53b82937b2220bfa13ca28d5eec95e9e24a4b3de74a89ad08d925088eacb8"} Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.792388 4945 generic.go:334] "Generic (PLEG): container finished" podID="a7bde639-beb8-4042-9adc-ad1dcde01cfc" containerID="93551780c8644f2f3dd32af635522d9a8fd7000bb4d63bc0667edbca9bf5877e" exitCode=0 Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.792425 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxqrl" event={"ID":"a7bde639-beb8-4042-9adc-ad1dcde01cfc","Type":"ContainerDied","Data":"93551780c8644f2f3dd32af635522d9a8fd7000bb4d63bc0667edbca9bf5877e"} Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.792449 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxqrl" event={"ID":"a7bde639-beb8-4042-9adc-ad1dcde01cfc","Type":"ContainerStarted","Data":"54ac2779ffacc777c08b23cc72d4b0f927f58562eaf1f02b36be3f794a730333"} Oct 08 15:19:00 crc kubenswrapper[4945]: I1008 15:19:00.816406 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:01 crc kubenswrapper[4945]: I1008 15:19:01.030943 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-64lj5"] Oct 08 15:19:01 crc kubenswrapper[4945]: W1008 15:19:01.039476 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff2444c_a4f4_4779_a11c_968c0ed65815.slice/crio-9f14a851ab47d24799dda4347804899485079dc7a160c0b6a941880768f08817 WatchSource:0}: Error finding container 9f14a851ab47d24799dda4347804899485079dc7a160c0b6a941880768f08817: Status 404 returned error can't find the container with id 9f14a851ab47d24799dda4347804899485079dc7a160c0b6a941880768f08817 Oct 08 15:19:01 crc kubenswrapper[4945]: I1008 15:19:01.798061 4945 generic.go:334] "Generic (PLEG): container finished" podID="6ff2444c-a4f4-4779-a11c-968c0ed65815" containerID="3bf2bc163c5586912e9faeda06321a01d1f583c31bd185a6ec4b1afa7bd6af25" exitCode=0 Oct 08 15:19:01 crc kubenswrapper[4945]: I1008 15:19:01.798134 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64lj5" event={"ID":"6ff2444c-a4f4-4779-a11c-968c0ed65815","Type":"ContainerDied","Data":"3bf2bc163c5586912e9faeda06321a01d1f583c31bd185a6ec4b1afa7bd6af25"} Oct 08 15:19:01 crc kubenswrapper[4945]: I1008 15:19:01.798652 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64lj5" event={"ID":"6ff2444c-a4f4-4779-a11c-968c0ed65815","Type":"ContainerStarted","Data":"9f14a851ab47d24799dda4347804899485079dc7a160c0b6a941880768f08817"} Oct 08 15:19:02 crc kubenswrapper[4945]: I1008 15:19:02.807137 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xdxb" event={"ID":"db0ecadd-b135-46e5-aa0e-92d952913a11","Type":"ContainerStarted","Data":"4d0c3d48151e3f1b252ba3a190799b301a08491635cbbfdcba534f1a0ccdca34"} Oct 08 15:19:07 crc kubenswrapper[4945]: I1008 15:19:07.413342 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:19:07 crc kubenswrapper[4945]: I1008 15:19:07.413901 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:19:07 crc kubenswrapper[4945]: I1008 15:19:07.475435 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:19:07 crc kubenswrapper[4945]: I1008 15:19:07.492446 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9xdxb" podStartSLOduration=7.585108228 podStartE2EDuration="10.492430436s" podCreationTimestamp="2025-10-08 15:18:57 +0000 UTC" firstStartedPulling="2025-10-08 15:18:58.765548221 +0000 UTC m=+348.119463122" lastFinishedPulling="2025-10-08 15:19:01.672870429 +0000 UTC m=+351.026785330" observedRunningTime="2025-10-08 15:19:02.829372995 +0000 UTC m=+352.183287896" watchObservedRunningTime="2025-10-08 15:19:07.492430436 +0000 UTC m=+356.846345337" Oct 08 15:19:07 crc kubenswrapper[4945]: I1008 15:19:07.877707 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9xdxb" Oct 08 15:19:10 crc kubenswrapper[4945]: I1008 15:19:10.845322 4945 generic.go:334] "Generic (PLEG): container finished" podID="6ff2444c-a4f4-4779-a11c-968c0ed65815" containerID="1b8e8101a13256cb5f20480ca4db6a0c3362811173ea8ba6b96a5fa5a951f84c" exitCode=0 Oct 08 15:19:10 crc kubenswrapper[4945]: I1008 15:19:10.845410 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64lj5" event={"ID":"6ff2444c-a4f4-4779-a11c-968c0ed65815","Type":"ContainerDied","Data":"1b8e8101a13256cb5f20480ca4db6a0c3362811173ea8ba6b96a5fa5a951f84c"} Oct 08 15:19:10 crc kubenswrapper[4945]: I1008 15:19:10.847368 4945 generic.go:334] "Generic (PLEG): container finished" podID="a7bde639-beb8-4042-9adc-ad1dcde01cfc" containerID="004d787e4bfe1c3f54eb882f07a9cafeacf63c9b4259fbdc7e24d6235f28811b" exitCode=0 Oct 08 15:19:10 crc kubenswrapper[4945]: I1008 15:19:10.847415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxqrl" event={"ID":"a7bde639-beb8-4042-9adc-ad1dcde01cfc","Type":"ContainerDied","Data":"004d787e4bfe1c3f54eb882f07a9cafeacf63c9b4259fbdc7e24d6235f28811b"} Oct 08 15:19:10 crc kubenswrapper[4945]: I1008 15:19:10.848698 4945 generic.go:334] "Generic (PLEG): container finished" podID="9c54859f-2831-49c8-88db-4b0c4451f3b0" containerID="cb4ebf83b6f89d2fbe413263776f18e3455a18ccd7ab73d1b3ce9297789dc411" exitCode=0 Oct 08 15:19:10 crc kubenswrapper[4945]: I1008 15:19:10.848721 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rzsp2" event={"ID":"9c54859f-2831-49c8-88db-4b0c4451f3b0","Type":"ContainerDied","Data":"cb4ebf83b6f89d2fbe413263776f18e3455a18ccd7ab73d1b3ce9297789dc411"} Oct 08 15:19:14 crc kubenswrapper[4945]: I1008 15:19:14.869028 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rzsp2" event={"ID":"9c54859f-2831-49c8-88db-4b0c4451f3b0","Type":"ContainerStarted","Data":"250875c709e4cd355640ddc70b89a4f692b6198c22f2817f5e885bfb01dfcf6a"} Oct 08 15:19:14 crc kubenswrapper[4945]: I1008 15:19:14.893644 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rzsp2" podStartSLOduration=2.126167687 podStartE2EDuration="16.893624352s" podCreationTimestamp="2025-10-08 15:18:58 +0000 UTC" firstStartedPulling="2025-10-08 15:18:59.772956215 +0000 UTC m=+349.126871116" lastFinishedPulling="2025-10-08 15:19:14.54041288 +0000 UTC m=+363.894327781" observedRunningTime="2025-10-08 15:19:14.891819767 +0000 UTC m=+364.245734668" watchObservedRunningTime="2025-10-08 15:19:14.893624352 +0000 UTC m=+364.247539253" Oct 08 15:19:15 crc kubenswrapper[4945]: I1008 15:19:15.876450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64lj5" event={"ID":"6ff2444c-a4f4-4779-a11c-968c0ed65815","Type":"ContainerStarted","Data":"cb0661443e8a58290db3b5db0a89655802061447d12713ec3088fde6ed7773f4"} Oct 08 15:19:15 crc kubenswrapper[4945]: I1008 15:19:15.878467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxqrl" event={"ID":"a7bde639-beb8-4042-9adc-ad1dcde01cfc","Type":"ContainerStarted","Data":"51c6f7359c8dc3ea4b848f301706fa9f2f87f1dbb39acf502cac8da58850edac"} Oct 08 15:19:15 crc kubenswrapper[4945]: I1008 15:19:15.911999 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-64lj5" podStartSLOduration=2.753061115 podStartE2EDuration="15.911975893s" podCreationTimestamp="2025-10-08 15:19:00 +0000 UTC" firstStartedPulling="2025-10-08 15:19:01.79989422 +0000 UTC m=+351.153809121" lastFinishedPulling="2025-10-08 15:19:14.958808998 +0000 UTC m=+364.312723899" observedRunningTime="2025-10-08 15:19:15.90816454 +0000 UTC m=+365.262079441" watchObservedRunningTime="2025-10-08 15:19:15.911975893 +0000 UTC m=+365.265890794" Oct 08 15:19:15 crc kubenswrapper[4945]: I1008 15:19:15.933824 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cxqrl" podStartSLOduration=2.890824647 podStartE2EDuration="16.933804438s" podCreationTimestamp="2025-10-08 15:18:59 +0000 UTC" firstStartedPulling="2025-10-08 15:19:00.794392663 +0000 UTC m=+350.148307564" lastFinishedPulling="2025-10-08 15:19:14.837372464 +0000 UTC m=+364.191287355" observedRunningTime="2025-10-08 15:19:15.929964954 +0000 UTC m=+365.283879855" watchObservedRunningTime="2025-10-08 15:19:15.933804438 +0000 UTC m=+365.287719339" Oct 08 15:19:18 crc kubenswrapper[4945]: I1008 15:19:18.427165 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:19:18 crc kubenswrapper[4945]: I1008 15:19:18.428257 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:19:18 crc kubenswrapper[4945]: I1008 15:19:18.475634 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:19:19 crc kubenswrapper[4945]: I1008 15:19:19.183874 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:19:19 crc kubenswrapper[4945]: I1008 15:19:19.184237 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:19:19 crc kubenswrapper[4945]: I1008 15:19:19.820657 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:19:19 crc kubenswrapper[4945]: I1008 15:19:19.820728 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:19:19 crc kubenswrapper[4945]: I1008 15:19:19.878156 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:19:19 crc kubenswrapper[4945]: I1008 15:19:19.945982 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rzsp2" Oct 08 15:19:19 crc kubenswrapper[4945]: I1008 15:19:19.946477 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cxqrl" Oct 08 15:19:20 crc kubenswrapper[4945]: I1008 15:19:20.817258 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:20 crc kubenswrapper[4945]: I1008 15:19:20.818586 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:20 crc kubenswrapper[4945]: I1008 15:19:20.851035 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:20 crc kubenswrapper[4945]: I1008 15:19:20.959501 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-64lj5" Oct 08 15:19:49 crc kubenswrapper[4945]: I1008 15:19:49.184201 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:19:49 crc kubenswrapper[4945]: I1008 15:19:49.184893 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:20:19 crc kubenswrapper[4945]: I1008 15:20:19.184397 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:20:19 crc kubenswrapper[4945]: I1008 15:20:19.185504 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:20:19 crc kubenswrapper[4945]: I1008 15:20:19.185589 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:20:19 crc kubenswrapper[4945]: I1008 15:20:19.186674 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6dba648876d9e856a999caed1abc8004128f384ec2a39cd78a8487b03680a491"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:20:19 crc kubenswrapper[4945]: I1008 15:20:19.186793 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://6dba648876d9e856a999caed1abc8004128f384ec2a39cd78a8487b03680a491" gracePeriod=600 Oct 08 15:20:20 crc kubenswrapper[4945]: I1008 15:20:20.222211 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="6dba648876d9e856a999caed1abc8004128f384ec2a39cd78a8487b03680a491" exitCode=0 Oct 08 15:20:20 crc kubenswrapper[4945]: I1008 15:20:20.222256 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"6dba648876d9e856a999caed1abc8004128f384ec2a39cd78a8487b03680a491"} Oct 08 15:20:20 crc kubenswrapper[4945]: I1008 15:20:20.222513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"92e8cf284bb0b21015533deb9be10a938603af77ea1c43c763cb1d3ffcf03af1"} Oct 08 15:20:20 crc kubenswrapper[4945]: I1008 15:20:20.222537 4945 scope.go:117] "RemoveContainer" containerID="51895b51d46eecde84d50e88702864bf738c65aecbf16940f9ee5b0b13271341" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.033603 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8q6qp"] Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.035277 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.052429 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8q6qp"] Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198523 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f9136df-2f6f-4ec7-8c80-34fff5828078-trusted-ca\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-registry-tls\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198651 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198696 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f9136df-2f6f-4ec7-8c80-34fff5828078-registry-certificates\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f9136df-2f6f-4ec7-8c80-34fff5828078-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198785 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f9136df-2f6f-4ec7-8c80-34fff5828078-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198801 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-bound-sa-token\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.198835 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c968p\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-kube-api-access-c968p\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.228672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.299903 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f9136df-2f6f-4ec7-8c80-34fff5828078-registry-certificates\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.299954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f9136df-2f6f-4ec7-8c80-34fff5828078-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.299985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f9136df-2f6f-4ec7-8c80-34fff5828078-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.300004 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-bound-sa-token\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.300060 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c968p\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-kube-api-access-c968p\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.300093 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f9136df-2f6f-4ec7-8c80-34fff5828078-trusted-ca\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.300154 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-registry-tls\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.300734 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f9136df-2f6f-4ec7-8c80-34fff5828078-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.301593 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f9136df-2f6f-4ec7-8c80-34fff5828078-trusted-ca\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.301858 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f9136df-2f6f-4ec7-8c80-34fff5828078-registry-certificates\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.305942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-registry-tls\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.308133 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f9136df-2f6f-4ec7-8c80-34fff5828078-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.319946 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-bound-sa-token\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.326860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c968p\" (UniqueName: \"kubernetes.io/projected/4f9136df-2f6f-4ec7-8c80-34fff5828078-kube-api-access-c968p\") pod \"image-registry-66df7c8f76-8q6qp\" (UID: \"4f9136df-2f6f-4ec7-8c80-34fff5828078\") " pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.351911 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.544486 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8q6qp"] Oct 08 15:21:28 crc kubenswrapper[4945]: W1008 15:21:28.549397 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f9136df_2f6f_4ec7_8c80_34fff5828078.slice/crio-95126edea52e76d936759724971312d7055431a1f9bd9422a3eff361c62416e7 WatchSource:0}: Error finding container 95126edea52e76d936759724971312d7055431a1f9bd9422a3eff361c62416e7: Status 404 returned error can't find the container with id 95126edea52e76d936759724971312d7055431a1f9bd9422a3eff361c62416e7 Oct 08 15:21:28 crc kubenswrapper[4945]: I1008 15:21:28.620399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" event={"ID":"4f9136df-2f6f-4ec7-8c80-34fff5828078","Type":"ContainerStarted","Data":"95126edea52e76d936759724971312d7055431a1f9bd9422a3eff361c62416e7"} Oct 08 15:21:29 crc kubenswrapper[4945]: I1008 15:21:29.628229 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" event={"ID":"4f9136df-2f6f-4ec7-8c80-34fff5828078","Type":"ContainerStarted","Data":"bb2da52eddc44daaca328081b8e620608395ad9299f102142c8634605e88ff7b"} Oct 08 15:21:29 crc kubenswrapper[4945]: I1008 15:21:29.628721 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:29 crc kubenswrapper[4945]: I1008 15:21:29.655905 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" podStartSLOduration=1.6558803370000001 podStartE2EDuration="1.655880337s" podCreationTimestamp="2025-10-08 15:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:21:29.650618047 +0000 UTC m=+499.004532988" watchObservedRunningTime="2025-10-08 15:21:29.655880337 +0000 UTC m=+499.009795278" Oct 08 15:21:48 crc kubenswrapper[4945]: I1008 15:21:48.360247 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8q6qp" Oct 08 15:21:48 crc kubenswrapper[4945]: I1008 15:21:48.420512 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-922gb"] Oct 08 15:22:12 crc kubenswrapper[4945]: I1008 15:22:12.153076 4945 scope.go:117] "RemoveContainer" containerID="390dffb6684ce85858d269f3a21f53acdbca17a88c20fcb80e5a9a056efff58e" Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.467424 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" podUID="5fec4a94-36ab-4892-a13c-35a21108d2ba" containerName="registry" containerID="cri-o://25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff" gracePeriod=30 Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.885507 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.890702 4945 generic.go:334] "Generic (PLEG): container finished" podID="5fec4a94-36ab-4892-a13c-35a21108d2ba" containerID="25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff" exitCode=0 Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.890753 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" event={"ID":"5fec4a94-36ab-4892-a13c-35a21108d2ba","Type":"ContainerDied","Data":"25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff"} Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.890781 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" event={"ID":"5fec4a94-36ab-4892-a13c-35a21108d2ba","Type":"ContainerDied","Data":"78b07e3004bdc8823f90dd1cbe6f71ffe7420ddc55ddaaea66c54d420abb67c6"} Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.890797 4945 scope.go:117] "RemoveContainer" containerID="25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff" Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.890958 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-922gb" Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.920726 4945 scope.go:117] "RemoveContainer" containerID="25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff" Oct 08 15:22:13 crc kubenswrapper[4945]: E1008 15:22:13.921519 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff\": container with ID starting with 25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff not found: ID does not exist" containerID="25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff" Oct 08 15:22:13 crc kubenswrapper[4945]: I1008 15:22:13.921579 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff"} err="failed to get container status \"25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff\": rpc error: code = NotFound desc = could not find container \"25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff\": container with ID starting with 25f978f2cac134731a4a7acaef014d289248a4a16294f1645119b6c9000d4aff not found: ID does not exist" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.022664 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5fec4a94-36ab-4892-a13c-35a21108d2ba-ca-trust-extracted\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.022785 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb7bl\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-kube-api-access-wb7bl\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.023090 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.023177 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-tls\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.023228 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5fec4a94-36ab-4892-a13c-35a21108d2ba-installation-pull-secrets\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.023283 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-bound-sa-token\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.023340 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-certificates\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.023391 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-trusted-ca\") pod \"5fec4a94-36ab-4892-a13c-35a21108d2ba\" (UID: \"5fec4a94-36ab-4892-a13c-35a21108d2ba\") " Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.024166 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.024621 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.029312 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fec4a94-36ab-4892-a13c-35a21108d2ba-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.029447 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.029998 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-kube-api-access-wb7bl" (OuterVolumeSpecName: "kube-api-access-wb7bl") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "kube-api-access-wb7bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.033809 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.033842 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.045759 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fec4a94-36ab-4892-a13c-35a21108d2ba-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "5fec4a94-36ab-4892-a13c-35a21108d2ba" (UID: "5fec4a94-36ab-4892-a13c-35a21108d2ba"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.125001 4945 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5fec4a94-36ab-4892-a13c-35a21108d2ba-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.125048 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb7bl\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-kube-api-access-wb7bl\") on node \"crc\" DevicePath \"\"" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.125066 4945 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.125078 4945 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5fec4a94-36ab-4892-a13c-35a21108d2ba-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.125090 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5fec4a94-36ab-4892-a13c-35a21108d2ba-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.125100 4945 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.125135 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5fec4a94-36ab-4892-a13c-35a21108d2ba-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.218843 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-922gb"] Oct 08 15:22:14 crc kubenswrapper[4945]: I1008 15:22:14.223919 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-922gb"] Oct 08 15:22:16 crc kubenswrapper[4945]: I1008 15:22:16.036237 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fec4a94-36ab-4892-a13c-35a21108d2ba" path="/var/lib/kubelet/pods/5fec4a94-36ab-4892-a13c-35a21108d2ba/volumes" Oct 08 15:22:49 crc kubenswrapper[4945]: I1008 15:22:49.184694 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:22:49 crc kubenswrapper[4945]: I1008 15:22:49.185417 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:23:19 crc kubenswrapper[4945]: I1008 15:23:19.184772 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:23:19 crc kubenswrapper[4945]: I1008 15:23:19.185358 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.183903 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.185293 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.185348 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.186288 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92e8cf284bb0b21015533deb9be10a938603af77ea1c43c763cb1d3ffcf03af1"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.186361 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://92e8cf284bb0b21015533deb9be10a938603af77ea1c43c763cb1d3ffcf03af1" gracePeriod=600 Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.455614 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="92e8cf284bb0b21015533deb9be10a938603af77ea1c43c763cb1d3ffcf03af1" exitCode=0 Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.455987 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"92e8cf284bb0b21015533deb9be10a938603af77ea1c43c763cb1d3ffcf03af1"} Oct 08 15:23:49 crc kubenswrapper[4945]: I1008 15:23:49.456086 4945 scope.go:117] "RemoveContainer" containerID="6dba648876d9e856a999caed1abc8004128f384ec2a39cd78a8487b03680a491" Oct 08 15:23:50 crc kubenswrapper[4945]: I1008 15:23:50.462844 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"ec80442af9d49173b57d5cd8aba7514d1075212c202e9f51e35b871dde889d11"} Oct 08 15:24:12 crc kubenswrapper[4945]: I1008 15:24:12.215804 4945 scope.go:117] "RemoveContainer" containerID="a2233bbbbb7656424294e68ad211ee54c65710eba68f45fa5ad87907c685ce5e" Oct 08 15:24:12 crc kubenswrapper[4945]: I1008 15:24:12.239087 4945 scope.go:117] "RemoveContainer" containerID="6a66b6e25ba18d6dbd13bc4f1b2026a695622989390fe4684de4e68b77040020" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.450085 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2pms8"] Oct 08 15:25:02 crc kubenswrapper[4945]: E1008 15:25:02.450921 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fec4a94-36ab-4892-a13c-35a21108d2ba" containerName="registry" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.450936 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fec4a94-36ab-4892-a13c-35a21108d2ba" containerName="registry" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.451068 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fec4a94-36ab-4892-a13c-35a21108d2ba" containerName="registry" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.451519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.453643 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-2mgxp" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.457904 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.458389 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.479752 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2pms8"] Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.485792 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xhl5p"] Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.486418 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xhl5p" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.492655 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-8rfss" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.496229 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-q9846"] Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.496975 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.507877 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-cknct" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.515311 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xhl5p"] Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.525645 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-q9846"] Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.544795 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb2p7\" (UniqueName: \"kubernetes.io/projected/a020faff-9bc7-4f4c-83a4-9e21f352e5f2-kube-api-access-sb2p7\") pod \"cert-manager-cainjector-7f985d654d-2pms8\" (UID: \"a020faff-9bc7-4f4c-83a4-9e21f352e5f2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.645378 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrwbf\" (UniqueName: \"kubernetes.io/projected/8a063d76-5a7a-49f7-b19c-fac4fd303a90-kube-api-access-mrwbf\") pod \"cert-manager-5b446d88c5-xhl5p\" (UID: \"8a063d76-5a7a-49f7-b19c-fac4fd303a90\") " pod="cert-manager/cert-manager-5b446d88c5-xhl5p" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.645426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk45t\" (UniqueName: \"kubernetes.io/projected/ec83ad94-ee5b-4984-8381-bda1289c4ef8-kube-api-access-lk45t\") pod \"cert-manager-webhook-5655c58dd6-q9846\" (UID: \"ec83ad94-ee5b-4984-8381-bda1289c4ef8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.645507 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb2p7\" (UniqueName: \"kubernetes.io/projected/a020faff-9bc7-4f4c-83a4-9e21f352e5f2-kube-api-access-sb2p7\") pod \"cert-manager-cainjector-7f985d654d-2pms8\" (UID: \"a020faff-9bc7-4f4c-83a4-9e21f352e5f2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.669978 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb2p7\" (UniqueName: \"kubernetes.io/projected/a020faff-9bc7-4f4c-83a4-9e21f352e5f2-kube-api-access-sb2p7\") pod \"cert-manager-cainjector-7f985d654d-2pms8\" (UID: \"a020faff-9bc7-4f4c-83a4-9e21f352e5f2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.746873 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrwbf\" (UniqueName: \"kubernetes.io/projected/8a063d76-5a7a-49f7-b19c-fac4fd303a90-kube-api-access-mrwbf\") pod \"cert-manager-5b446d88c5-xhl5p\" (UID: \"8a063d76-5a7a-49f7-b19c-fac4fd303a90\") " pod="cert-manager/cert-manager-5b446d88c5-xhl5p" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.746929 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk45t\" (UniqueName: \"kubernetes.io/projected/ec83ad94-ee5b-4984-8381-bda1289c4ef8-kube-api-access-lk45t\") pod \"cert-manager-webhook-5655c58dd6-q9846\" (UID: \"ec83ad94-ee5b-4984-8381-bda1289c4ef8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.762358 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrwbf\" (UniqueName: \"kubernetes.io/projected/8a063d76-5a7a-49f7-b19c-fac4fd303a90-kube-api-access-mrwbf\") pod \"cert-manager-5b446d88c5-xhl5p\" (UID: \"8a063d76-5a7a-49f7-b19c-fac4fd303a90\") " pod="cert-manager/cert-manager-5b446d88c5-xhl5p" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.765905 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk45t\" (UniqueName: \"kubernetes.io/projected/ec83ad94-ee5b-4984-8381-bda1289c4ef8-kube-api-access-lk45t\") pod \"cert-manager-webhook-5655c58dd6-q9846\" (UID: \"ec83ad94-ee5b-4984-8381-bda1289c4ef8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.768539 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.811811 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xhl5p" Oct 08 15:25:02 crc kubenswrapper[4945]: I1008 15:25:02.833970 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" Oct 08 15:25:03 crc kubenswrapper[4945]: I1008 15:25:03.003258 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2pms8"] Oct 08 15:25:03 crc kubenswrapper[4945]: I1008 15:25:03.031167 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:25:03 crc kubenswrapper[4945]: I1008 15:25:03.318150 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-q9846"] Oct 08 15:25:03 crc kubenswrapper[4945]: I1008 15:25:03.323307 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xhl5p"] Oct 08 15:25:03 crc kubenswrapper[4945]: W1008 15:25:03.331331 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a063d76_5a7a_49f7_b19c_fac4fd303a90.slice/crio-6c3315b84d245cb36c5a7664161cf1b9bc12cd4eca68ba5b34caf614604410a5 WatchSource:0}: Error finding container 6c3315b84d245cb36c5a7664161cf1b9bc12cd4eca68ba5b34caf614604410a5: Status 404 returned error can't find the container with id 6c3315b84d245cb36c5a7664161cf1b9bc12cd4eca68ba5b34caf614604410a5 Oct 08 15:25:03 crc kubenswrapper[4945]: I1008 15:25:03.878864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xhl5p" event={"ID":"8a063d76-5a7a-49f7-b19c-fac4fd303a90","Type":"ContainerStarted","Data":"6c3315b84d245cb36c5a7664161cf1b9bc12cd4eca68ba5b34caf614604410a5"} Oct 08 15:25:03 crc kubenswrapper[4945]: I1008 15:25:03.879864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" event={"ID":"a020faff-9bc7-4f4c-83a4-9e21f352e5f2","Type":"ContainerStarted","Data":"153c0a525b991b1626204128c5f77235c51a16ee5845c86d215d34f9f71e8433"} Oct 08 15:25:03 crc kubenswrapper[4945]: I1008 15:25:03.880936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" event={"ID":"ec83ad94-ee5b-4984-8381-bda1289c4ef8","Type":"ContainerStarted","Data":"1690b25702cd45bd712ca21b40d58b32151874db4eadb35608acb23a4ecb67f6"} Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.855144 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9g2vg"] Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.856029 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-controller" containerID="cri-o://0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" gracePeriod=30 Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.856479 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="sbdb" containerID="cri-o://4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" gracePeriod=30 Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.856522 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="nbdb" containerID="cri-o://312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" gracePeriod=30 Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.856567 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="northd" containerID="cri-o://795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" gracePeriod=30 Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.856613 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" gracePeriod=30 Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.856664 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-node" containerID="cri-o://db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" gracePeriod=30 Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.856708 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-acl-logging" containerID="cri-o://8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" gracePeriod=30 Oct 08 15:25:12 crc kubenswrapper[4945]: I1008 15:25:12.890677 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" containerID="cri-o://4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" gracePeriod=30 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.619137 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/3.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.621674 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovn-acl-logging/0.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.622257 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovn-controller/0.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.622724 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674518 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lwwgs"] Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674744 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="nbdb" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674758 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="nbdb" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674773 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674781 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674792 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-node" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674800 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-node" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674809 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674816 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674832 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674840 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674849 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kubecfg-setup" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674856 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kubecfg-setup" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674868 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="sbdb" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674874 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="sbdb" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674886 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-acl-logging" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674893 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-acl-logging" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674903 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674910 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674919 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="northd" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674926 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="northd" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.674936 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.674944 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675048 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-node" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675061 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="nbdb" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675070 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675126 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675139 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675152 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-acl-logging" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675163 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675171 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovn-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675182 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="northd" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675191 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="sbdb" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.675300 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675309 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.675321 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675329 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675446 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.675643 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" containerName="ovnkube-controller" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.679796 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.787907 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-netns\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.787938 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-slash\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.787958 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-ovn\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.787971 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-node-log\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.787991 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-script-lib\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788011 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-kubelet\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788034 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28239584-598a-49d2-a9b0-189e4f4ad733-ovn-node-metrics-cert\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788060 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-ovn-kubernetes\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788089 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-var-lib-openvswitch\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788148 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-config\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788180 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-env-overrides\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788201 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fzlp\" (UniqueName: \"kubernetes.io/projected/28239584-598a-49d2-a9b0-189e4f4ad733-kube-api-access-8fzlp\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788231 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-etc-openvswitch\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788258 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-netd\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788282 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-var-lib-cni-networks-ovn-kubernetes\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788311 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-log-socket\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788329 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-openvswitch\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788341 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-systemd\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788358 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-systemd-units\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-bin\") pod \"28239584-598a-49d2-a9b0-189e4f4ad733\" (UID: \"28239584-598a-49d2-a9b0-189e4f4ad733\") " Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788506 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovnkube-script-lib\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788536 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-ovn\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788560 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-node-log\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-run-ovn-kubernetes\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788600 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788628 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-kubelet\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788642 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mfg5\" (UniqueName: \"kubernetes.io/projected/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-kube-api-access-4mfg5\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788665 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-systemd\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788680 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-var-lib-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788696 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovn-node-metrics-cert\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788718 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-systemd-units\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788738 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-env-overrides\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788794 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-run-netns\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-etc-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788842 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovnkube-config\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-cni-bin\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788891 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788913 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-log-socket\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788936 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-slash\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.788954 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-cni-netd\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.789566 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.789575 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.789768 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-log-socket" (OuterVolumeSpecName: "log-socket") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.789723 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.789737 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.789750 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.789813 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.790003 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-slash" (OuterVolumeSpecName: "host-slash") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.790119 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.790231 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-node-log" (OuterVolumeSpecName: "node-log") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.790868 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.790973 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.791095 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.791475 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.791449 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.791569 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.791579 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.798218 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28239584-598a-49d2-a9b0-189e4f4ad733-kube-api-access-8fzlp" (OuterVolumeSpecName: "kube-api-access-8fzlp") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "kube-api-access-8fzlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.809631 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28239584-598a-49d2-a9b0-189e4f4ad733-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.822995 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "28239584-598a-49d2-a9b0-189e4f4ad733" (UID: "28239584-598a-49d2-a9b0-189e4f4ad733"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889666 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889719 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-kubelet\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889744 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mfg5\" (UniqueName: \"kubernetes.io/projected/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-kube-api-access-4mfg5\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889772 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-var-lib-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889789 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-systemd\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889805 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovn-node-metrics-cert\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889830 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-systemd-units\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889846 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-run-netns\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889861 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-env-overrides\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889877 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-etc-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889892 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovnkube-config\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-cni-bin\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890342 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890411 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-log-socket\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890427 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-slash\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890467 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-cni-netd\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889967 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-systemd-units\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890707 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-etc-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889876 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-kubelet\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890775 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-cni-bin\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889931 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890837 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890883 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-slash\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890925 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-log-socket\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.889999 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-var-lib-openvswitch\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-systemd\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890993 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-cni-netd\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-env-overrides\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.890094 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-run-netns\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891581 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovnkube-script-lib\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891603 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-ovn\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891644 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-node-log\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891659 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-run-ovn-kubernetes\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891724 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891735 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fzlp\" (UniqueName: \"kubernetes.io/projected/28239584-598a-49d2-a9b0-189e4f4ad733-kube-api-access-8fzlp\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891744 4945 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891752 4945 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891761 4945 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891770 4945 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-log-socket\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891801 4945 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891809 4945 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891817 4945 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891825 4945 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891832 4945 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891840 4945 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-slash\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891849 4945 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891890 4945 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-node-log\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891898 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891905 4945 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891914 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28239584-598a-49d2-a9b0-189e4f4ad733-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891922 4945 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891930 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28239584-598a-49d2-a9b0-189e4f4ad733-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891960 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28239584-598a-49d2-a9b0-189e4f4ad733-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.891986 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-host-run-ovn-kubernetes\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.892314 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-run-ovn\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.892378 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-node-log\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.892676 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovnkube-script-lib\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.894581 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovn-node-metrics-cert\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.898418 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-ovnkube-config\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.907751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mfg5\" (UniqueName: \"kubernetes.io/projected/b2653be1-1d7b-4a9e-a751-3ceea2eddeb4-kube-api-access-4mfg5\") pod \"ovnkube-node-lwwgs\" (UID: \"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.934250 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" event={"ID":"a020faff-9bc7-4f4c-83a4-9e21f352e5f2","Type":"ContainerStarted","Data":"6b4e1c2577fe9ab58395484b3b6e6bfdba61cdeff63e709898217a862b968c49"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.936060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" event={"ID":"ec83ad94-ee5b-4984-8381-bda1289c4ef8","Type":"ContainerStarted","Data":"dbf234afa7cf6f8f2cece8e1456d9c4045b5821c517b408af3312e0d403c147d"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.936154 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.937226 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xhl5p" event={"ID":"8a063d76-5a7a-49f7-b19c-fac4fd303a90","Type":"ContainerStarted","Data":"5199c871e608e0e22352538b1b5e75a5cbe4b2043c05802d40b7e0774acb5850"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.939029 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovnkube-controller/3.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.941342 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovn-acl-logging/0.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.941946 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9g2vg_28239584-598a-49d2-a9b0-189e4f4ad733/ovn-controller/0.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942440 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" exitCode=0 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942470 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" exitCode=0 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942481 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" exitCode=0 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942490 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" exitCode=0 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942499 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" exitCode=0 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942508 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" exitCode=0 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942516 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" exitCode=143 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942526 4945 generic.go:334] "Generic (PLEG): container finished" podID="28239584-598a-49d2-a9b0-189e4f4ad733" containerID="0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" exitCode=143 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942576 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942610 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942633 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942650 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942677 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942691 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942707 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942718 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942728 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942737 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942748 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942757 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942767 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942774 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942785 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942799 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942810 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942820 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942830 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942840 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942851 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942861 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942864 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942884 4945 scope.go:117] "RemoveContainer" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.942872 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943564 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943620 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943750 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943806 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943851 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943896 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.943956 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.944008 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.944054 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.944760 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.944883 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.944978 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945063 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9g2vg" event={"ID":"28239584-598a-49d2-a9b0-189e4f4ad733","Type":"ContainerDied","Data":"6ec74449b25519e4606575db149042f9c896c0632effe38a06596c77fc39fa9b"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945206 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945305 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945376 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945428 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945490 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945542 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945594 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945643 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945691 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.945740 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.947538 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/2.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.947927 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-2pms8" podStartSLOduration=1.2852908570000001 podStartE2EDuration="11.947914855s" podCreationTimestamp="2025-10-08 15:25:02 +0000 UTC" firstStartedPulling="2025-10-08 15:25:03.030881083 +0000 UTC m=+712.384795994" lastFinishedPulling="2025-10-08 15:25:13.693505081 +0000 UTC m=+723.047419992" observedRunningTime="2025-10-08 15:25:13.946493639 +0000 UTC m=+723.300408540" watchObservedRunningTime="2025-10-08 15:25:13.947914855 +0000 UTC m=+723.301829766" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.947971 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/1.log" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.948147 4945 generic.go:334] "Generic (PLEG): container finished" podID="22e38fb8-8588-494f-a6cb-8b3d2c3a3142" containerID="e4e3fedefec003a21b24bf0cb252dbec6df8b7b9ba1b1d3af62aadd0be7e8bcd" exitCode=2 Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.948175 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerDied","Data":"e4e3fedefec003a21b24bf0cb252dbec6df8b7b9ba1b1d3af62aadd0be7e8bcd"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.948260 4945 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203"} Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.949733 4945 scope.go:117] "RemoveContainer" containerID="e4e3fedefec003a21b24bf0cb252dbec6df8b7b9ba1b1d3af62aadd0be7e8bcd" Oct 08 15:25:13 crc kubenswrapper[4945]: E1008 15:25:13.950484 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-j8kc9_openshift-multus(22e38fb8-8588-494f-a6cb-8b3d2c3a3142)\"" pod="openshift-multus/multus-j8kc9" podUID="22e38fb8-8588-494f-a6cb-8b3d2c3a3142" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.963541 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.967723 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" podStartSLOduration=1.685476169 podStartE2EDuration="11.967703717s" podCreationTimestamp="2025-10-08 15:25:02 +0000 UTC" firstStartedPulling="2025-10-08 15:25:03.3251976 +0000 UTC m=+712.679112521" lastFinishedPulling="2025-10-08 15:25:13.607425168 +0000 UTC m=+722.961340069" observedRunningTime="2025-10-08 15:25:13.963712959 +0000 UTC m=+723.317627870" watchObservedRunningTime="2025-10-08 15:25:13.967703717 +0000 UTC m=+723.321618618" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.979130 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-xhl5p" podStartSLOduration=1.703678414 podStartE2EDuration="11.979089781s" podCreationTimestamp="2025-10-08 15:25:02 +0000 UTC" firstStartedPulling="2025-10-08 15:25:03.334147144 +0000 UTC m=+712.688062065" lastFinishedPulling="2025-10-08 15:25:13.609558531 +0000 UTC m=+722.963473432" observedRunningTime="2025-10-08 15:25:13.975314187 +0000 UTC m=+723.329229088" watchObservedRunningTime="2025-10-08 15:25:13.979089781 +0000 UTC m=+723.333004682" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.983951 4945 scope.go:117] "RemoveContainer" containerID="4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" Oct 08 15:25:13 crc kubenswrapper[4945]: I1008 15:25:13.996043 4945 scope.go:117] "RemoveContainer" containerID="312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.016530 4945 scope.go:117] "RemoveContainer" containerID="795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.019783 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9g2vg"] Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.023156 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9g2vg"] Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.035425 4945 scope.go:117] "RemoveContainer" containerID="c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.039014 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28239584-598a-49d2-a9b0-189e4f4ad733" path="/var/lib/kubelet/pods/28239584-598a-49d2-a9b0-189e4f4ad733/volumes" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.065774 4945 scope.go:117] "RemoveContainer" containerID="db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.078928 4945 scope.go:117] "RemoveContainer" containerID="8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.087545 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.090837 4945 scope.go:117] "RemoveContainer" containerID="0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.103310 4945 scope.go:117] "RemoveContainer" containerID="7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d" Oct 08 15:25:14 crc kubenswrapper[4945]: W1008 15:25:14.109532 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2653be1_1d7b_4a9e_a751_3ceea2eddeb4.slice/crio-e32732734821543c8a58b203e41f81c8b081ff7a62902a405d5565a4af2d3e1c WatchSource:0}: Error finding container e32732734821543c8a58b203e41f81c8b081ff7a62902a405d5565a4af2d3e1c: Status 404 returned error can't find the container with id e32732734821543c8a58b203e41f81c8b081ff7a62902a405d5565a4af2d3e1c Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.138874 4945 scope.go:117] "RemoveContainer" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.139501 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": container with ID starting with 4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634 not found: ID does not exist" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.139554 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} err="failed to get container status \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": rpc error: code = NotFound desc = could not find container \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": container with ID starting with 4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.139589 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.140023 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": container with ID starting with cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511 not found: ID does not exist" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.141151 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} err="failed to get container status \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": rpc error: code = NotFound desc = could not find container \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": container with ID starting with cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.141254 4945 scope.go:117] "RemoveContainer" containerID="4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.141630 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": container with ID starting with 4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345 not found: ID does not exist" containerID="4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.141689 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} err="failed to get container status \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": rpc error: code = NotFound desc = could not find container \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": container with ID starting with 4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.141725 4945 scope.go:117] "RemoveContainer" containerID="312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.142618 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": container with ID starting with 312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a not found: ID does not exist" containerID="312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.142711 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} err="failed to get container status \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": rpc error: code = NotFound desc = could not find container \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": container with ID starting with 312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.142780 4945 scope.go:117] "RemoveContainer" containerID="795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.143077 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": container with ID starting with 795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4 not found: ID does not exist" containerID="795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.143176 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} err="failed to get container status \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": rpc error: code = NotFound desc = could not find container \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": container with ID starting with 795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.143249 4945 scope.go:117] "RemoveContainer" containerID="c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.143537 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": container with ID starting with c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e not found: ID does not exist" containerID="c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.143563 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} err="failed to get container status \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": rpc error: code = NotFound desc = could not find container \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": container with ID starting with c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.143584 4945 scope.go:117] "RemoveContainer" containerID="db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.143849 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": container with ID starting with db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3 not found: ID does not exist" containerID="db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.143878 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} err="failed to get container status \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": rpc error: code = NotFound desc = could not find container \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": container with ID starting with db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.143895 4945 scope.go:117] "RemoveContainer" containerID="8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.144237 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": container with ID starting with 8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc not found: ID does not exist" containerID="8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.144335 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} err="failed to get container status \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": rpc error: code = NotFound desc = could not find container \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": container with ID starting with 8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.144418 4945 scope.go:117] "RemoveContainer" containerID="0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.144725 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": container with ID starting with 0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56 not found: ID does not exist" containerID="0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.144814 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} err="failed to get container status \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": rpc error: code = NotFound desc = could not find container \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": container with ID starting with 0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.144919 4945 scope.go:117] "RemoveContainer" containerID="7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d" Oct 08 15:25:14 crc kubenswrapper[4945]: E1008 15:25:14.145281 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": container with ID starting with 7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d not found: ID does not exist" containerID="7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.145370 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} err="failed to get container status \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": rpc error: code = NotFound desc = could not find container \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": container with ID starting with 7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.145461 4945 scope.go:117] "RemoveContainer" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.145870 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} err="failed to get container status \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": rpc error: code = NotFound desc = could not find container \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": container with ID starting with 4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.145958 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.146292 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} err="failed to get container status \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": rpc error: code = NotFound desc = could not find container \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": container with ID starting with cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.146396 4945 scope.go:117] "RemoveContainer" containerID="4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.146731 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} err="failed to get container status \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": rpc error: code = NotFound desc = could not find container \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": container with ID starting with 4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.146823 4945 scope.go:117] "RemoveContainer" containerID="312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.147138 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} err="failed to get container status \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": rpc error: code = NotFound desc = could not find container \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": container with ID starting with 312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.147239 4945 scope.go:117] "RemoveContainer" containerID="795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.147533 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} err="failed to get container status \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": rpc error: code = NotFound desc = could not find container \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": container with ID starting with 795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.147565 4945 scope.go:117] "RemoveContainer" containerID="c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.147824 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} err="failed to get container status \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": rpc error: code = NotFound desc = could not find container \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": container with ID starting with c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.147896 4945 scope.go:117] "RemoveContainer" containerID="db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.148197 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} err="failed to get container status \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": rpc error: code = NotFound desc = could not find container \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": container with ID starting with db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.148279 4945 scope.go:117] "RemoveContainer" containerID="8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.148557 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} err="failed to get container status \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": rpc error: code = NotFound desc = could not find container \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": container with ID starting with 8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.148636 4945 scope.go:117] "RemoveContainer" containerID="0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.148929 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} err="failed to get container status \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": rpc error: code = NotFound desc = could not find container \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": container with ID starting with 0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.149035 4945 scope.go:117] "RemoveContainer" containerID="7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.149311 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} err="failed to get container status \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": rpc error: code = NotFound desc = could not find container \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": container with ID starting with 7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.149400 4945 scope.go:117] "RemoveContainer" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.149666 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} err="failed to get container status \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": rpc error: code = NotFound desc = could not find container \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": container with ID starting with 4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.149743 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.150018 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} err="failed to get container status \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": rpc error: code = NotFound desc = could not find container \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": container with ID starting with cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.150137 4945 scope.go:117] "RemoveContainer" containerID="4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.150498 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} err="failed to get container status \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": rpc error: code = NotFound desc = could not find container \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": container with ID starting with 4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.150580 4945 scope.go:117] "RemoveContainer" containerID="312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.150888 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} err="failed to get container status \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": rpc error: code = NotFound desc = could not find container \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": container with ID starting with 312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.150982 4945 scope.go:117] "RemoveContainer" containerID="795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.151285 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} err="failed to get container status \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": rpc error: code = NotFound desc = could not find container \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": container with ID starting with 795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.151368 4945 scope.go:117] "RemoveContainer" containerID="c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.151584 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} err="failed to get container status \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": rpc error: code = NotFound desc = could not find container \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": container with ID starting with c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.151657 4945 scope.go:117] "RemoveContainer" containerID="db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.151942 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} err="failed to get container status \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": rpc error: code = NotFound desc = could not find container \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": container with ID starting with db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.152015 4945 scope.go:117] "RemoveContainer" containerID="8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.152220 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} err="failed to get container status \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": rpc error: code = NotFound desc = could not find container \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": container with ID starting with 8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.152292 4945 scope.go:117] "RemoveContainer" containerID="0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.152575 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} err="failed to get container status \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": rpc error: code = NotFound desc = could not find container \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": container with ID starting with 0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.152656 4945 scope.go:117] "RemoveContainer" containerID="7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.152976 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} err="failed to get container status \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": rpc error: code = NotFound desc = could not find container \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": container with ID starting with 7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.153051 4945 scope.go:117] "RemoveContainer" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.153330 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} err="failed to get container status \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": rpc error: code = NotFound desc = could not find container \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": container with ID starting with 4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.153400 4945 scope.go:117] "RemoveContainer" containerID="cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.153601 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511"} err="failed to get container status \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": rpc error: code = NotFound desc = could not find container \"cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511\": container with ID starting with cc25e5c0e2323296400f593830a5a540b014876180b345c6c0ac47d7462bf511 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.153680 4945 scope.go:117] "RemoveContainer" containerID="4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.153928 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345"} err="failed to get container status \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": rpc error: code = NotFound desc = could not find container \"4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345\": container with ID starting with 4e5165e41651ce022cd36ddbf59a429ecef4f3c8c9448760c2ee33a40cb3f345 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.154001 4945 scope.go:117] "RemoveContainer" containerID="312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.154298 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a"} err="failed to get container status \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": rpc error: code = NotFound desc = could not find container \"312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a\": container with ID starting with 312c940947f6b98b43a88983f4cd72291128087ab4505721f7f6db956ff8a10a not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.154379 4945 scope.go:117] "RemoveContainer" containerID="795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.154636 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4"} err="failed to get container status \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": rpc error: code = NotFound desc = could not find container \"795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4\": container with ID starting with 795e9daf9239a2199e6a933f6babd1a09d4d81c56780ef5ede4865a3709232a4 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.154713 4945 scope.go:117] "RemoveContainer" containerID="c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.154936 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e"} err="failed to get container status \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": rpc error: code = NotFound desc = could not find container \"c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e\": container with ID starting with c9aea555f8fa0d615f1fdecca82439a1ec24244fe1012bcc3e6ad988a296788e not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.155011 4945 scope.go:117] "RemoveContainer" containerID="db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.155287 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3"} err="failed to get container status \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": rpc error: code = NotFound desc = could not find container \"db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3\": container with ID starting with db2afb184e415ee5095921de1a846fef97da5c02e61be14b980ecccae49cf5b3 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.155376 4945 scope.go:117] "RemoveContainer" containerID="8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.155720 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc"} err="failed to get container status \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": rpc error: code = NotFound desc = could not find container \"8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc\": container with ID starting with 8345c5083b245adda819b2aea90bcd3b8b45a98e5c4bf1cb75bd1a51e72dcecc not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.155808 4945 scope.go:117] "RemoveContainer" containerID="0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.156187 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56"} err="failed to get container status \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": rpc error: code = NotFound desc = could not find container \"0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56\": container with ID starting with 0c4e0ad4f74a81b80504f0ec9371d21d07a1d3040bbfe91ef9bbacd168fa5e56 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.156276 4945 scope.go:117] "RemoveContainer" containerID="7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.156516 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d"} err="failed to get container status \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": rpc error: code = NotFound desc = could not find container \"7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d\": container with ID starting with 7be6d4e4631693ded018ad261a4eeb5534fee985d7935a84e0e6a462dbf5323d not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.156592 4945 scope.go:117] "RemoveContainer" containerID="4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.156852 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634"} err="failed to get container status \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": rpc error: code = NotFound desc = could not find container \"4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634\": container with ID starting with 4ebc28464e057f643eeeff972de73dcbf37a15036e4b1e2ba91dbd5315d15634 not found: ID does not exist" Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.955856 4945 generic.go:334] "Generic (PLEG): container finished" podID="b2653be1-1d7b-4a9e-a751-3ceea2eddeb4" containerID="68552008012b362d93f86fb6b86c9f0204a14b43c22b476d91fd80938315ffd9" exitCode=0 Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.955998 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerDied","Data":"68552008012b362d93f86fb6b86c9f0204a14b43c22b476d91fd80938315ffd9"} Oct 08 15:25:14 crc kubenswrapper[4945]: I1008 15:25:14.956047 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"e32732734821543c8a58b203e41f81c8b081ff7a62902a405d5565a4af2d3e1c"} Oct 08 15:25:15 crc kubenswrapper[4945]: I1008 15:25:15.965312 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"0582ee946229aab201476501fa6ea43cdecf971bb7e3951d5c17bd00f9567c09"} Oct 08 15:25:15 crc kubenswrapper[4945]: I1008 15:25:15.965650 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"2384f77592fdbc16299f85576f3db30769a95a317260f72122944746a1215ba7"} Oct 08 15:25:15 crc kubenswrapper[4945]: I1008 15:25:15.965662 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"ba65d0147320d405f7c1d2d0a0a6235fdab870a967453a48746f751bd914fd9e"} Oct 08 15:25:15 crc kubenswrapper[4945]: I1008 15:25:15.965671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"5d8f1aba778903d6626a23b85a60a040630b81fd55e29043f3bd4a5a2b05a3e2"} Oct 08 15:25:15 crc kubenswrapper[4945]: I1008 15:25:15.965679 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"c9fae47c79dd7df2459beb9043bffbc1e4adb6a191049819635973dee01b9d3c"} Oct 08 15:25:15 crc kubenswrapper[4945]: I1008 15:25:15.965687 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"17763f7f785c19e1f8858e7193001c3cc72ebcab0475d2658c1c26cf1540be3e"} Oct 08 15:25:18 crc kubenswrapper[4945]: I1008 15:25:18.985188 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"4ce2d74696aa22b6b93e48c20b390831bb508d7be77b164e77a7f25c9a4a0895"} Oct 08 15:25:20 crc kubenswrapper[4945]: I1008 15:25:20.999506 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" event={"ID":"b2653be1-1d7b-4a9e-a751-3ceea2eddeb4","Type":"ContainerStarted","Data":"e994ee979efff3a11216f2f95b252de60c190eeb6179e87b64cd955d68e588b2"} Oct 08 15:25:20 crc kubenswrapper[4945]: I1008 15:25:20.999923 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:21 crc kubenswrapper[4945]: I1008 15:25:20.999946 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:21 crc kubenswrapper[4945]: I1008 15:25:21.024356 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:21 crc kubenswrapper[4945]: I1008 15:25:21.040533 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" podStartSLOduration=8.040505554 podStartE2EDuration="8.040505554s" podCreationTimestamp="2025-10-08 15:25:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:25:21.040426072 +0000 UTC m=+730.394340983" watchObservedRunningTime="2025-10-08 15:25:21.040505554 +0000 UTC m=+730.394420455" Oct 08 15:25:22 crc kubenswrapper[4945]: I1008 15:25:22.004538 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:22 crc kubenswrapper[4945]: I1008 15:25:22.045283 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:22 crc kubenswrapper[4945]: I1008 15:25:22.837274 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-q9846" Oct 08 15:25:27 crc kubenswrapper[4945]: I1008 15:25:27.024182 4945 scope.go:117] "RemoveContainer" containerID="e4e3fedefec003a21b24bf0cb252dbec6df8b7b9ba1b1d3af62aadd0be7e8bcd" Oct 08 15:25:27 crc kubenswrapper[4945]: E1008 15:25:27.024816 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-j8kc9_openshift-multus(22e38fb8-8588-494f-a6cb-8b3d2c3a3142)\"" pod="openshift-multus/multus-j8kc9" podUID="22e38fb8-8588-494f-a6cb-8b3d2c3a3142" Oct 08 15:25:42 crc kubenswrapper[4945]: I1008 15:25:42.030613 4945 scope.go:117] "RemoveContainer" containerID="e4e3fedefec003a21b24bf0cb252dbec6df8b7b9ba1b1d3af62aadd0be7e8bcd" Oct 08 15:25:43 crc kubenswrapper[4945]: I1008 15:25:43.134833 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/2.log" Oct 08 15:25:43 crc kubenswrapper[4945]: I1008 15:25:43.135612 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/1.log" Oct 08 15:25:43 crc kubenswrapper[4945]: I1008 15:25:43.135656 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8kc9" event={"ID":"22e38fb8-8588-494f-a6cb-8b3d2c3a3142","Type":"ContainerStarted","Data":"8d77127e0dd0a09c6c848b15944568c31ba2240606a408c3f9ae24dfe953f601"} Oct 08 15:25:44 crc kubenswrapper[4945]: I1008 15:25:44.110766 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lwwgs" Oct 08 15:25:49 crc kubenswrapper[4945]: I1008 15:25:49.184068 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:25:49 crc kubenswrapper[4945]: I1008 15:25:49.184676 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.789276 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks"] Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.790595 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.792901 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.800890 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks"] Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.869806 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjhrt\" (UniqueName: \"kubernetes.io/projected/81f62690-bcd9-4d49-ac08-c754f28e143b-kube-api-access-zjhrt\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.870217 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.870374 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.971881 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.972336 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjhrt\" (UniqueName: \"kubernetes.io/projected/81f62690-bcd9-4d49-ac08-c754f28e143b-kube-api-access-zjhrt\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.972493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.972666 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.972873 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:51 crc kubenswrapper[4945]: I1008 15:25:51.998056 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjhrt\" (UniqueName: \"kubernetes.io/projected/81f62690-bcd9-4d49-ac08-c754f28e143b-kube-api-access-zjhrt\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:52 crc kubenswrapper[4945]: I1008 15:25:52.106783 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:25:52 crc kubenswrapper[4945]: I1008 15:25:52.524586 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks"] Oct 08 15:25:52 crc kubenswrapper[4945]: I1008 15:25:52.867395 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xghvq"] Oct 08 15:25:52 crc kubenswrapper[4945]: I1008 15:25:52.867689 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" podUID="b5f113bd-0207-40ce-9793-c7b0c901b885" containerName="controller-manager" containerID="cri-o://53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc" gracePeriod=30 Oct 08 15:25:52 crc kubenswrapper[4945]: I1008 15:25:52.973360 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg"] Oct 08 15:25:52 crc kubenswrapper[4945]: I1008 15:25:52.973611 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" podUID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" containerName="route-controller-manager" containerID="cri-o://96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37" gracePeriod=30 Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.194924 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" event={"ID":"81f62690-bcd9-4d49-ac08-c754f28e143b","Type":"ContainerStarted","Data":"74a08c4bcac2ea36465977c2b6af078bc8de0c05241d8a618ce39a42291c140d"} Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.195269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" event={"ID":"81f62690-bcd9-4d49-ac08-c754f28e143b","Type":"ContainerStarted","Data":"3e523148b2cb794c48a297d0565ac5b2e8d4e67dc098e1940fad8e720f82ab04"} Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.758717 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.903897 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-client-ca\") pod \"b5f113bd-0207-40ce-9793-c7b0c901b885\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.903944 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-config\") pod \"b5f113bd-0207-40ce-9793-c7b0c901b885\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.903966 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvn2n\" (UniqueName: \"kubernetes.io/projected/b5f113bd-0207-40ce-9793-c7b0c901b885-kube-api-access-cvn2n\") pod \"b5f113bd-0207-40ce-9793-c7b0c901b885\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.904023 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f113bd-0207-40ce-9793-c7b0c901b885-serving-cert\") pod \"b5f113bd-0207-40ce-9793-c7b0c901b885\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.904048 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-proxy-ca-bundles\") pod \"b5f113bd-0207-40ce-9793-c7b0c901b885\" (UID: \"b5f113bd-0207-40ce-9793-c7b0c901b885\") " Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.904645 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-client-ca" (OuterVolumeSpecName: "client-ca") pod "b5f113bd-0207-40ce-9793-c7b0c901b885" (UID: "b5f113bd-0207-40ce-9793-c7b0c901b885"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.904691 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-config" (OuterVolumeSpecName: "config") pod "b5f113bd-0207-40ce-9793-c7b0c901b885" (UID: "b5f113bd-0207-40ce-9793-c7b0c901b885"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.904801 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b5f113bd-0207-40ce-9793-c7b0c901b885" (UID: "b5f113bd-0207-40ce-9793-c7b0c901b885"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.913577 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5f113bd-0207-40ce-9793-c7b0c901b885-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b5f113bd-0207-40ce-9793-c7b0c901b885" (UID: "b5f113bd-0207-40ce-9793-c7b0c901b885"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.914096 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5f113bd-0207-40ce-9793-c7b0c901b885-kube-api-access-cvn2n" (OuterVolumeSpecName: "kube-api-access-cvn2n") pod "b5f113bd-0207-40ce-9793-c7b0c901b885" (UID: "b5f113bd-0207-40ce-9793-c7b0c901b885"). InnerVolumeSpecName "kube-api-access-cvn2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:25:53 crc kubenswrapper[4945]: I1008 15:25:53.947444 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.010773 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.010814 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.010827 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvn2n\" (UniqueName: \"kubernetes.io/projected/b5f113bd-0207-40ce-9793-c7b0c901b885-kube-api-access-cvn2n\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.010840 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f113bd-0207-40ce-9793-c7b0c901b885-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.010851 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5f113bd-0207-40ce-9793-c7b0c901b885-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.112199 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjkb6\" (UniqueName: \"kubernetes.io/projected/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-kube-api-access-mjkb6\") pod \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.112282 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-config\") pod \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.112339 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-client-ca\") pod \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.112420 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-serving-cert\") pod \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\" (UID: \"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193\") " Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.113136 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-config" (OuterVolumeSpecName: "config") pod "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" (UID: "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.113129 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-client-ca" (OuterVolumeSpecName: "client-ca") pod "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" (UID: "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.117355 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" (UID: "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.129187 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-kube-api-access-mjkb6" (OuterVolumeSpecName: "kube-api-access-mjkb6") pod "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" (UID: "500dbefa-9b5b-4dcf-8cc2-a70b4fa35193"). InnerVolumeSpecName "kube-api-access-mjkb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.201144 4945 generic.go:334] "Generic (PLEG): container finished" podID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerID="74a08c4bcac2ea36465977c2b6af078bc8de0c05241d8a618ce39a42291c140d" exitCode=0 Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.201214 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" event={"ID":"81f62690-bcd9-4d49-ac08-c754f28e143b","Type":"ContainerDied","Data":"74a08c4bcac2ea36465977c2b6af078bc8de0c05241d8a618ce39a42291c140d"} Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.203623 4945 generic.go:334] "Generic (PLEG): container finished" podID="b5f113bd-0207-40ce-9793-c7b0c901b885" containerID="53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc" exitCode=0 Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.203694 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" event={"ID":"b5f113bd-0207-40ce-9793-c7b0c901b885","Type":"ContainerDied","Data":"53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc"} Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.203705 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.203721 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xghvq" event={"ID":"b5f113bd-0207-40ce-9793-c7b0c901b885","Type":"ContainerDied","Data":"48a002ba0ed1591f492bfdbfe64365da649bb12f1789f7247b952f391e33811e"} Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.203736 4945 scope.go:117] "RemoveContainer" containerID="53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.210891 4945 generic.go:334] "Generic (PLEG): container finished" podID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" containerID="96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37" exitCode=0 Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.210930 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" event={"ID":"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193","Type":"ContainerDied","Data":"96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37"} Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.210954 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" event={"ID":"500dbefa-9b5b-4dcf-8cc2-a70b4fa35193","Type":"ContainerDied","Data":"bd9bf508c258f0143c49eb8de9769c51d86f272e2a4ef694e59113d4ab8aebea"} Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.210997 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.213861 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.213892 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.213904 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.213915 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjkb6\" (UniqueName: \"kubernetes.io/projected/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193-kube-api-access-mjkb6\") on node \"crc\" DevicePath \"\"" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.226551 4945 scope.go:117] "RemoveContainer" containerID="53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc" Oct 08 15:25:54 crc kubenswrapper[4945]: E1008 15:25:54.227695 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc\": container with ID starting with 53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc not found: ID does not exist" containerID="53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.227723 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc"} err="failed to get container status \"53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc\": rpc error: code = NotFound desc = could not find container \"53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc\": container with ID starting with 53c07bb2ea0539df693f50832d2c83514593e330c50a188c504189cca99209fc not found: ID does not exist" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.227740 4945 scope.go:117] "RemoveContainer" containerID="96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.242173 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xghvq"] Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.244375 4945 scope.go:117] "RemoveContainer" containerID="96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37" Oct 08 15:25:54 crc kubenswrapper[4945]: E1008 15:25:54.245285 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37\": container with ID starting with 96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37 not found: ID does not exist" containerID="96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.245406 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37"} err="failed to get container status \"96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37\": rpc error: code = NotFound desc = could not find container \"96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37\": container with ID starting with 96d72b64e3e24471d5f1ec3cd358602a23088f6536653076d78f658354bc7b37 not found: ID does not exist" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.256598 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xghvq"] Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.274799 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg"] Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.280364 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-477zg"] Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.408602 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46"] Oct 08 15:25:54 crc kubenswrapper[4945]: E1008 15:25:54.408947 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f113bd-0207-40ce-9793-c7b0c901b885" containerName="controller-manager" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.408968 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f113bd-0207-40ce-9793-c7b0c901b885" containerName="controller-manager" Oct 08 15:25:54 crc kubenswrapper[4945]: E1008 15:25:54.408990 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" containerName="route-controller-manager" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.408999 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" containerName="route-controller-manager" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.409180 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5f113bd-0207-40ce-9793-c7b0c901b885" containerName="controller-manager" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.409206 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" containerName="route-controller-manager" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.409614 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.411651 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d84d98779-w5cpp"] Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.412481 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.413453 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.413511 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.413469 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.413785 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.416290 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.416677 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.416886 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.417572 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.417779 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.420053 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.420322 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.421756 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.425221 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.425414 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d84d98779-w5cpp"] Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.430043 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46"] Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525008 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-proxy-ca-bundles\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525495 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-client-ca\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525551 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8786b584-de03-4708-8a76-2eb84f3e7cf6-client-ca\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525610 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-config\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786b584-de03-4708-8a76-2eb84f3e7cf6-config\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8786b584-de03-4708-8a76-2eb84f3e7cf6-serving-cert\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525792 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq25n\" (UniqueName: \"kubernetes.io/projected/3ef1d4a6-4a16-4261-9625-fadbcd067c59-kube-api-access-pq25n\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525816 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdnm9\" (UniqueName: \"kubernetes.io/projected/8786b584-de03-4708-8a76-2eb84f3e7cf6-kube-api-access-bdnm9\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.525833 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ef1d4a6-4a16-4261-9625-fadbcd067c59-serving-cert\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.626935 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-proxy-ca-bundles\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627003 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-client-ca\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627023 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8786b584-de03-4708-8a76-2eb84f3e7cf6-client-ca\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627046 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-config\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627076 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786b584-de03-4708-8a76-2eb84f3e7cf6-config\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627094 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8786b584-de03-4708-8a76-2eb84f3e7cf6-serving-cert\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq25n\" (UniqueName: \"kubernetes.io/projected/3ef1d4a6-4a16-4261-9625-fadbcd067c59-kube-api-access-pq25n\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdnm9\" (UniqueName: \"kubernetes.io/projected/8786b584-de03-4708-8a76-2eb84f3e7cf6-kube-api-access-bdnm9\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.627165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ef1d4a6-4a16-4261-9625-fadbcd067c59-serving-cert\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.628051 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-client-ca\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.628100 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8786b584-de03-4708-8a76-2eb84f3e7cf6-client-ca\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.628491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-proxy-ca-bundles\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.628958 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1d4a6-4a16-4261-9625-fadbcd067c59-config\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.629917 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786b584-de03-4708-8a76-2eb84f3e7cf6-config\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.632382 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8786b584-de03-4708-8a76-2eb84f3e7cf6-serving-cert\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.632414 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ef1d4a6-4a16-4261-9625-fadbcd067c59-serving-cert\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.642608 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq25n\" (UniqueName: \"kubernetes.io/projected/3ef1d4a6-4a16-4261-9625-fadbcd067c59-kube-api-access-pq25n\") pod \"controller-manager-5d84d98779-w5cpp\" (UID: \"3ef1d4a6-4a16-4261-9625-fadbcd067c59\") " pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.649069 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdnm9\" (UniqueName: \"kubernetes.io/projected/8786b584-de03-4708-8a76-2eb84f3e7cf6-kube-api-access-bdnm9\") pod \"route-controller-manager-f894cfff4-xxs46\" (UID: \"8786b584-de03-4708-8a76-2eb84f3e7cf6\") " pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.736341 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.744168 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:54 crc kubenswrapper[4945]: I1008 15:25:54.986300 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46"] Oct 08 15:25:55 crc kubenswrapper[4945]: I1008 15:25:55.016302 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d84d98779-w5cpp"] Oct 08 15:25:55 crc kubenswrapper[4945]: W1008 15:25:55.024562 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ef1d4a6_4a16_4261_9625_fadbcd067c59.slice/crio-a7bcded36db22058d9d16502e1fa541f5af18f1c8763ce3830491f8dec837d5d WatchSource:0}: Error finding container a7bcded36db22058d9d16502e1fa541f5af18f1c8763ce3830491f8dec837d5d: Status 404 returned error can't find the container with id a7bcded36db22058d9d16502e1fa541f5af18f1c8763ce3830491f8dec837d5d Oct 08 15:25:55 crc kubenswrapper[4945]: I1008 15:25:55.215777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" event={"ID":"8786b584-de03-4708-8a76-2eb84f3e7cf6","Type":"ContainerStarted","Data":"b663faf782c072cdf9966c2189dcf3d535f5072f4431ebc4f0b9c756b250bec0"} Oct 08 15:25:55 crc kubenswrapper[4945]: I1008 15:25:55.217901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" event={"ID":"3ef1d4a6-4a16-4261-9625-fadbcd067c59","Type":"ContainerStarted","Data":"a7bcded36db22058d9d16502e1fa541f5af18f1c8763ce3830491f8dec837d5d"} Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.035514 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="500dbefa-9b5b-4dcf-8cc2-a70b4fa35193" path="/var/lib/kubelet/pods/500dbefa-9b5b-4dcf-8cc2-a70b4fa35193/volumes" Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.036764 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5f113bd-0207-40ce-9793-c7b0c901b885" path="/var/lib/kubelet/pods/b5f113bd-0207-40ce-9793-c7b0c901b885/volumes" Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.226425 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" event={"ID":"8786b584-de03-4708-8a76-2eb84f3e7cf6","Type":"ContainerStarted","Data":"9b2a7c37ddb2423530f2cd4e7971d32257c70780e20bcc43fe0659dab8b3ead2"} Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.226902 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.228592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" event={"ID":"3ef1d4a6-4a16-4261-9625-fadbcd067c59","Type":"ContainerStarted","Data":"b1f10e003302fce1761e22e48d8236bbac6af53086859e871dd707c4b273f466"} Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.228821 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.234105 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.236711 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.260085 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-f894cfff4-xxs46" podStartSLOduration=4.260062148 podStartE2EDuration="4.260062148s" podCreationTimestamp="2025-10-08 15:25:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:25:56.25368288 +0000 UTC m=+765.607597811" watchObservedRunningTime="2025-10-08 15:25:56.260062148 +0000 UTC m=+765.613977079" Oct 08 15:25:56 crc kubenswrapper[4945]: I1008 15:25:56.307128 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d84d98779-w5cpp" podStartSLOduration=3.307078599 podStartE2EDuration="3.307078599s" podCreationTimestamp="2025-10-08 15:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:25:56.302255248 +0000 UTC m=+765.656170189" watchObservedRunningTime="2025-10-08 15:25:56.307078599 +0000 UTC m=+765.660993520" Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.248811 4945 generic.go:334] "Generic (PLEG): container finished" podID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerID="a9d48b4b5b3eb4d8ed45aa43f9197b6bb5689c8c65e91f0af1ac9e1fca0ba54d" exitCode=0 Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.248880 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" event={"ID":"81f62690-bcd9-4d49-ac08-c754f28e143b","Type":"ContainerDied","Data":"a9d48b4b5b3eb4d8ed45aa43f9197b6bb5689c8c65e91f0af1ac9e1fca0ba54d"} Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.816913 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qdq9c"] Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.818685 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.831593 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qdq9c"] Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.901941 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-catalog-content\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.902023 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-utilities\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:25:59 crc kubenswrapper[4945]: I1008 15:25:59.902059 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k42jw\" (UniqueName: \"kubernetes.io/projected/f4d0e76f-95c0-465f-b059-188802063bac-kube-api-access-k42jw\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.003193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-utilities\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.003278 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k42jw\" (UniqueName: \"kubernetes.io/projected/f4d0e76f-95c0-465f-b059-188802063bac-kube-api-access-k42jw\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.003370 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-catalog-content\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.003897 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-utilities\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.004466 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-catalog-content\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.026119 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k42jw\" (UniqueName: \"kubernetes.io/projected/f4d0e76f-95c0-465f-b059-188802063bac-kube-api-access-k42jw\") pod \"redhat-operators-qdq9c\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.197766 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.264279 4945 generic.go:334] "Generic (PLEG): container finished" podID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerID="8fe4442fda4ca60bd2618bede89495c2fba5b993b7011390358b4ceec4904455" exitCode=0 Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.264313 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" event={"ID":"81f62690-bcd9-4d49-ac08-c754f28e143b","Type":"ContainerDied","Data":"8fe4442fda4ca60bd2618bede89495c2fba5b993b7011390358b4ceec4904455"} Oct 08 15:26:00 crc kubenswrapper[4945]: I1008 15:26:00.455693 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qdq9c"] Oct 08 15:26:00 crc kubenswrapper[4945]: W1008 15:26:00.468325 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4d0e76f_95c0_465f_b059_188802063bac.slice/crio-4e4532a0418b109469e7f6246fef22925747e7aa8d1fd1aabca69dc1cb5b92f3 WatchSource:0}: Error finding container 4e4532a0418b109469e7f6246fef22925747e7aa8d1fd1aabca69dc1cb5b92f3: Status 404 returned error can't find the container with id 4e4532a0418b109469e7f6246fef22925747e7aa8d1fd1aabca69dc1cb5b92f3 Oct 08 15:26:00 crc kubenswrapper[4945]: E1008 15:26:00.955225 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4d0e76f_95c0_465f_b059_188802063bac.slice/crio-b58a18e3c9fa070fb4412123ece15312fc95d9e3a55f44f6aa8f397554ee6f8d.scope\": RecentStats: unable to find data in memory cache]" Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.280403 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4d0e76f-95c0-465f-b059-188802063bac" containerID="b58a18e3c9fa070fb4412123ece15312fc95d9e3a55f44f6aa8f397554ee6f8d" exitCode=0 Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.281340 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdq9c" event={"ID":"f4d0e76f-95c0-465f-b059-188802063bac","Type":"ContainerDied","Data":"b58a18e3c9fa070fb4412123ece15312fc95d9e3a55f44f6aa8f397554ee6f8d"} Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.281478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdq9c" event={"ID":"f4d0e76f-95c0-465f-b059-188802063bac","Type":"ContainerStarted","Data":"4e4532a0418b109469e7f6246fef22925747e7aa8d1fd1aabca69dc1cb5b92f3"} Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.615902 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.725563 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjhrt\" (UniqueName: \"kubernetes.io/projected/81f62690-bcd9-4d49-ac08-c754f28e143b-kube-api-access-zjhrt\") pod \"81f62690-bcd9-4d49-ac08-c754f28e143b\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.725634 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-util\") pod \"81f62690-bcd9-4d49-ac08-c754f28e143b\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.725682 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-bundle\") pod \"81f62690-bcd9-4d49-ac08-c754f28e143b\" (UID: \"81f62690-bcd9-4d49-ac08-c754f28e143b\") " Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.727771 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-bundle" (OuterVolumeSpecName: "bundle") pod "81f62690-bcd9-4d49-ac08-c754f28e143b" (UID: "81f62690-bcd9-4d49-ac08-c754f28e143b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.731336 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f62690-bcd9-4d49-ac08-c754f28e143b-kube-api-access-zjhrt" (OuterVolumeSpecName: "kube-api-access-zjhrt") pod "81f62690-bcd9-4d49-ac08-c754f28e143b" (UID: "81f62690-bcd9-4d49-ac08-c754f28e143b"). InnerVolumeSpecName "kube-api-access-zjhrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.735827 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-util" (OuterVolumeSpecName: "util") pod "81f62690-bcd9-4d49-ac08-c754f28e143b" (UID: "81f62690-bcd9-4d49-ac08-c754f28e143b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.826642 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjhrt\" (UniqueName: \"kubernetes.io/projected/81f62690-bcd9-4d49-ac08-c754f28e143b-kube-api-access-zjhrt\") on node \"crc\" DevicePath \"\"" Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.826937 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-util\") on node \"crc\" DevicePath \"\"" Oct 08 15:26:01 crc kubenswrapper[4945]: I1008 15:26:01.826951 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/81f62690-bcd9-4d49-ac08-c754f28e143b-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:26:02 crc kubenswrapper[4945]: I1008 15:26:02.290178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" event={"ID":"81f62690-bcd9-4d49-ac08-c754f28e143b","Type":"ContainerDied","Data":"3e523148b2cb794c48a297d0565ac5b2e8d4e67dc098e1940fad8e720f82ab04"} Oct 08 15:26:02 crc kubenswrapper[4945]: I1008 15:26:02.290234 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e523148b2cb794c48a297d0565ac5b2e8d4e67dc098e1940fad8e720f82ab04" Oct 08 15:26:02 crc kubenswrapper[4945]: I1008 15:26:02.290243 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks" Oct 08 15:26:02 crc kubenswrapper[4945]: I1008 15:26:02.747585 4945 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 15:26:03 crc kubenswrapper[4945]: I1008 15:26:03.301268 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdq9c" event={"ID":"f4d0e76f-95c0-465f-b059-188802063bac","Type":"ContainerStarted","Data":"1f97ee03f05bd2e983bb2c289a62311383cd354a37e3fbf29b0f986239e26b5c"} Oct 08 15:26:04 crc kubenswrapper[4945]: I1008 15:26:04.310278 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4d0e76f-95c0-465f-b059-188802063bac" containerID="1f97ee03f05bd2e983bb2c289a62311383cd354a37e3fbf29b0f986239e26b5c" exitCode=0 Oct 08 15:26:04 crc kubenswrapper[4945]: I1008 15:26:04.310396 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdq9c" event={"ID":"f4d0e76f-95c0-465f-b059-188802063bac","Type":"ContainerDied","Data":"1f97ee03f05bd2e983bb2c289a62311383cd354a37e3fbf29b0f986239e26b5c"} Oct 08 15:26:05 crc kubenswrapper[4945]: I1008 15:26:05.323919 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdq9c" event={"ID":"f4d0e76f-95c0-465f-b059-188802063bac","Type":"ContainerStarted","Data":"33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225"} Oct 08 15:26:05 crc kubenswrapper[4945]: I1008 15:26:05.352046 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qdq9c" podStartSLOduration=2.617886388 podStartE2EDuration="6.352025033s" podCreationTimestamp="2025-10-08 15:25:59 +0000 UTC" firstStartedPulling="2025-10-08 15:26:01.28821414 +0000 UTC m=+770.642129051" lastFinishedPulling="2025-10-08 15:26:05.022352785 +0000 UTC m=+774.376267696" observedRunningTime="2025-10-08 15:26:05.349941097 +0000 UTC m=+774.703856038" watchObservedRunningTime="2025-10-08 15:26:05.352025033 +0000 UTC m=+774.705939964" Oct 08 15:26:10 crc kubenswrapper[4945]: I1008 15:26:10.198208 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:10 crc kubenswrapper[4945]: I1008 15:26:10.199758 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:11 crc kubenswrapper[4945]: I1008 15:26:11.262189 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qdq9c" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="registry-server" probeResult="failure" output=< Oct 08 15:26:11 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 15:26:11 crc kubenswrapper[4945]: > Oct 08 15:26:12 crc kubenswrapper[4945]: I1008 15:26:12.277759 4945 scope.go:117] "RemoveContainer" containerID="892c88fd98daeb518c0767497ac882dc99e93c965d8be3f7a207fa8b2539a203" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.363742 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8kc9_22e38fb8-8588-494f-a6cb-8b3d2c3a3142/kube-multus/2.log" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.777644 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k"] Oct 08 15:26:13 crc kubenswrapper[4945]: E1008 15:26:13.777874 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerName="extract" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.777886 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerName="extract" Oct 08 15:26:13 crc kubenswrapper[4945]: E1008 15:26:13.777895 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerName="util" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.777903 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerName="util" Oct 08 15:26:13 crc kubenswrapper[4945]: E1008 15:26:13.777921 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerName="pull" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.777928 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerName="pull" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.778035 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f62690-bcd9-4d49-ac08-c754f28e143b" containerName="extract" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.778493 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.780914 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-mwc89" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.781652 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.783303 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.793195 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k"] Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.878103 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sllk\" (UniqueName: \"kubernetes.io/projected/18169f3c-d245-4b63-8ea7-d6d3bd70cce4-kube-api-access-8sllk\") pod \"obo-prometheus-operator-7c8cf85677-g4w8k\" (UID: \"18169f3c-d245-4b63-8ea7-d6d3bd70cce4\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.903021 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd"] Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.903661 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.906040 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.906416 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-t45kn" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.916804 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd"] Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.920065 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp"] Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.920881 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.957473 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp"] Oct 08 15:26:13 crc kubenswrapper[4945]: I1008 15:26:13.979386 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sllk\" (UniqueName: \"kubernetes.io/projected/18169f3c-d245-4b63-8ea7-d6d3bd70cce4-kube-api-access-8sllk\") pod \"obo-prometheus-operator-7c8cf85677-g4w8k\" (UID: \"18169f3c-d245-4b63-8ea7-d6d3bd70cce4\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.002186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sllk\" (UniqueName: \"kubernetes.io/projected/18169f3c-d245-4b63-8ea7-d6d3bd70cce4-kube-api-access-8sllk\") pod \"obo-prometheus-operator-7c8cf85677-g4w8k\" (UID: \"18169f3c-d245-4b63-8ea7-d6d3bd70cce4\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.080866 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fed6f952-f438-4a3b-bd8c-f2832a1c6e8d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd\" (UID: \"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.080967 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/658dd3b7-2709-4785-aa43-64f041d46b7a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp\" (UID: \"658dd3b7-2709-4785-aa43-64f041d46b7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.081012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/658dd3b7-2709-4785-aa43-64f041d46b7a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp\" (UID: \"658dd3b7-2709-4785-aa43-64f041d46b7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.081078 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fed6f952-f438-4a3b-bd8c-f2832a1c6e8d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd\" (UID: \"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.094588 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.107217 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qrx7m"] Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.108007 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.110442 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-vjz4s" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.110605 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.115974 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qrx7m"] Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.182732 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/658dd3b7-2709-4785-aa43-64f041d46b7a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp\" (UID: \"658dd3b7-2709-4785-aa43-64f041d46b7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.182786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/658dd3b7-2709-4785-aa43-64f041d46b7a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp\" (UID: \"658dd3b7-2709-4785-aa43-64f041d46b7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.182852 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fed6f952-f438-4a3b-bd8c-f2832a1c6e8d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd\" (UID: \"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.182911 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fed6f952-f438-4a3b-bd8c-f2832a1c6e8d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd\" (UID: \"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.186547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fed6f952-f438-4a3b-bd8c-f2832a1c6e8d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd\" (UID: \"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.188459 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fed6f952-f438-4a3b-bd8c-f2832a1c6e8d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd\" (UID: \"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.188622 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/658dd3b7-2709-4785-aa43-64f041d46b7a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp\" (UID: \"658dd3b7-2709-4785-aa43-64f041d46b7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.189714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/658dd3b7-2709-4785-aa43-64f041d46b7a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp\" (UID: \"658dd3b7-2709-4785-aa43-64f041d46b7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.204345 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-sw8w6"] Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.205061 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.213288 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-84hvb" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.215975 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.231424 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-sw8w6"] Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.242644 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.283811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/09a37473-34c8-43ff-a839-3f712f3d0605-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qrx7m\" (UID: \"09a37473-34c8-43ff-a839-3f712f3d0605\") " pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.284268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjpbt\" (UniqueName: \"kubernetes.io/projected/09a37473-34c8-43ff-a839-3f712f3d0605-kube-api-access-cjpbt\") pod \"observability-operator-cc5f78dfc-qrx7m\" (UID: \"09a37473-34c8-43ff-a839-3f712f3d0605\") " pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.386137 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjpbt\" (UniqueName: \"kubernetes.io/projected/09a37473-34c8-43ff-a839-3f712f3d0605-kube-api-access-cjpbt\") pod \"observability-operator-cc5f78dfc-qrx7m\" (UID: \"09a37473-34c8-43ff-a839-3f712f3d0605\") " pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.386213 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d39d5733-b343-4a33-866b-220a8191eda9-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-sw8w6\" (UID: \"d39d5733-b343-4a33-866b-220a8191eda9\") " pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.386281 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/09a37473-34c8-43ff-a839-3f712f3d0605-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qrx7m\" (UID: \"09a37473-34c8-43ff-a839-3f712f3d0605\") " pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.386316 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmr6k\" (UniqueName: \"kubernetes.io/projected/d39d5733-b343-4a33-866b-220a8191eda9-kube-api-access-jmr6k\") pod \"perses-operator-54bc95c9fb-sw8w6\" (UID: \"d39d5733-b343-4a33-866b-220a8191eda9\") " pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.403976 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/09a37473-34c8-43ff-a839-3f712f3d0605-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qrx7m\" (UID: \"09a37473-34c8-43ff-a839-3f712f3d0605\") " pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.411986 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjpbt\" (UniqueName: \"kubernetes.io/projected/09a37473-34c8-43ff-a839-3f712f3d0605-kube-api-access-cjpbt\") pod \"observability-operator-cc5f78dfc-qrx7m\" (UID: \"09a37473-34c8-43ff-a839-3f712f3d0605\") " pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.461463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.487802 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d39d5733-b343-4a33-866b-220a8191eda9-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-sw8w6\" (UID: \"d39d5733-b343-4a33-866b-220a8191eda9\") " pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.487902 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmr6k\" (UniqueName: \"kubernetes.io/projected/d39d5733-b343-4a33-866b-220a8191eda9-kube-api-access-jmr6k\") pod \"perses-operator-54bc95c9fb-sw8w6\" (UID: \"d39d5733-b343-4a33-866b-220a8191eda9\") " pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.489083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d39d5733-b343-4a33-866b-220a8191eda9-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-sw8w6\" (UID: \"d39d5733-b343-4a33-866b-220a8191eda9\") " pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.506892 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmr6k\" (UniqueName: \"kubernetes.io/projected/d39d5733-b343-4a33-866b-220a8191eda9-kube-api-access-jmr6k\") pod \"perses-operator-54bc95c9fb-sw8w6\" (UID: \"d39d5733-b343-4a33-866b-220a8191eda9\") " pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.531442 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.705137 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k"] Oct 08 15:26:14 crc kubenswrapper[4945]: W1008 15:26:14.719949 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18169f3c_d245_4b63_8ea7_d6d3bd70cce4.slice/crio-efb10f673db38d395e4997db0f11dfd13ef4d4c99d461f0b75eb62d9ef3eeaac WatchSource:0}: Error finding container efb10f673db38d395e4997db0f11dfd13ef4d4c99d461f0b75eb62d9ef3eeaac: Status 404 returned error can't find the container with id efb10f673db38d395e4997db0f11dfd13ef4d4c99d461f0b75eb62d9ef3eeaac Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.802941 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd"] Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.815600 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp"] Oct 08 15:26:14 crc kubenswrapper[4945]: W1008 15:26:14.825680 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod658dd3b7_2709_4785_aa43_64f041d46b7a.slice/crio-fda00af36ab5511d79fe3da2693bf79dda20711ef16c1c7edfc0e01568fe5d0d WatchSource:0}: Error finding container fda00af36ab5511d79fe3da2693bf79dda20711ef16c1c7edfc0e01568fe5d0d: Status 404 returned error can't find the container with id fda00af36ab5511d79fe3da2693bf79dda20711ef16c1c7edfc0e01568fe5d0d Oct 08 15:26:14 crc kubenswrapper[4945]: I1008 15:26:14.956755 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qrx7m"] Oct 08 15:26:14 crc kubenswrapper[4945]: W1008 15:26:14.968958 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09a37473_34c8_43ff_a839_3f712f3d0605.slice/crio-d37c5a01386e7c03ddeaf81985b3a002df65f6789ddd639b367ab5276a4b6343 WatchSource:0}: Error finding container d37c5a01386e7c03ddeaf81985b3a002df65f6789ddd639b367ab5276a4b6343: Status 404 returned error can't find the container with id d37c5a01386e7c03ddeaf81985b3a002df65f6789ddd639b367ab5276a4b6343 Oct 08 15:26:15 crc kubenswrapper[4945]: I1008 15:26:15.082216 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-sw8w6"] Oct 08 15:26:15 crc kubenswrapper[4945]: W1008 15:26:15.095094 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd39d5733_b343_4a33_866b_220a8191eda9.slice/crio-acd50ded89d450feffae7eae3f4ca8319a7a39fc3a7ec44878bc3320cce38454 WatchSource:0}: Error finding container acd50ded89d450feffae7eae3f4ca8319a7a39fc3a7ec44878bc3320cce38454: Status 404 returned error can't find the container with id acd50ded89d450feffae7eae3f4ca8319a7a39fc3a7ec44878bc3320cce38454 Oct 08 15:26:15 crc kubenswrapper[4945]: I1008 15:26:15.377152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" event={"ID":"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d","Type":"ContainerStarted","Data":"9bdbe5ac16c79fca1339e456a449c705e4af90d1cabd84cb8d7262e411f89be5"} Oct 08 15:26:15 crc kubenswrapper[4945]: I1008 15:26:15.378535 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" event={"ID":"658dd3b7-2709-4785-aa43-64f041d46b7a","Type":"ContainerStarted","Data":"fda00af36ab5511d79fe3da2693bf79dda20711ef16c1c7edfc0e01568fe5d0d"} Oct 08 15:26:15 crc kubenswrapper[4945]: I1008 15:26:15.380097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" event={"ID":"18169f3c-d245-4b63-8ea7-d6d3bd70cce4","Type":"ContainerStarted","Data":"efb10f673db38d395e4997db0f11dfd13ef4d4c99d461f0b75eb62d9ef3eeaac"} Oct 08 15:26:15 crc kubenswrapper[4945]: I1008 15:26:15.381513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" event={"ID":"d39d5733-b343-4a33-866b-220a8191eda9","Type":"ContainerStarted","Data":"acd50ded89d450feffae7eae3f4ca8319a7a39fc3a7ec44878bc3320cce38454"} Oct 08 15:26:15 crc kubenswrapper[4945]: I1008 15:26:15.382876 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" event={"ID":"09a37473-34c8-43ff-a839-3f712f3d0605","Type":"ContainerStarted","Data":"d37c5a01386e7c03ddeaf81985b3a002df65f6789ddd639b367ab5276a4b6343"} Oct 08 15:26:19 crc kubenswrapper[4945]: I1008 15:26:19.184761 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:26:19 crc kubenswrapper[4945]: I1008 15:26:19.185062 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:26:20 crc kubenswrapper[4945]: I1008 15:26:20.243679 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:20 crc kubenswrapper[4945]: I1008 15:26:20.300037 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:20 crc kubenswrapper[4945]: I1008 15:26:20.471567 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qdq9c"] Oct 08 15:26:21 crc kubenswrapper[4945]: I1008 15:26:21.418153 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qdq9c" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="registry-server" containerID="cri-o://33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225" gracePeriod=2 Oct 08 15:26:22 crc kubenswrapper[4945]: I1008 15:26:22.425408 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4d0e76f-95c0-465f-b059-188802063bac" containerID="33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225" exitCode=0 Oct 08 15:26:22 crc kubenswrapper[4945]: I1008 15:26:22.425464 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdq9c" event={"ID":"f4d0e76f-95c0-465f-b059-188802063bac","Type":"ContainerDied","Data":"33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225"} Oct 08 15:26:30 crc kubenswrapper[4945]: E1008 15:26:30.198437 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225 is running failed: container process not found" containerID="33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:26:30 crc kubenswrapper[4945]: E1008 15:26:30.200012 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225 is running failed: container process not found" containerID="33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:26:30 crc kubenswrapper[4945]: E1008 15:26:30.200502 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225 is running failed: container process not found" containerID="33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 15:26:30 crc kubenswrapper[4945]: E1008 15:26:30.200594 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-qdq9c" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="registry-server" Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.693522 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.707677 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k42jw\" (UniqueName: \"kubernetes.io/projected/f4d0e76f-95c0-465f-b059-188802063bac-kube-api-access-k42jw\") pod \"f4d0e76f-95c0-465f-b059-188802063bac\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.707748 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-utilities\") pod \"f4d0e76f-95c0-465f-b059-188802063bac\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.711619 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-utilities" (OuterVolumeSpecName: "utilities") pod "f4d0e76f-95c0-465f-b059-188802063bac" (UID: "f4d0e76f-95c0-465f-b059-188802063bac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.713225 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4d0e76f-95c0-465f-b059-188802063bac-kube-api-access-k42jw" (OuterVolumeSpecName: "kube-api-access-k42jw") pod "f4d0e76f-95c0-465f-b059-188802063bac" (UID: "f4d0e76f-95c0-465f-b059-188802063bac"). InnerVolumeSpecName "kube-api-access-k42jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.808688 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-catalog-content\") pod \"f4d0e76f-95c0-465f-b059-188802063bac\" (UID: \"f4d0e76f-95c0-465f-b059-188802063bac\") " Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.808994 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.809020 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k42jw\" (UniqueName: \"kubernetes.io/projected/f4d0e76f-95c0-465f-b059-188802063bac-kube-api-access-k42jw\") on node \"crc\" DevicePath \"\"" Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.902737 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4d0e76f-95c0-465f-b059-188802063bac" (UID: "f4d0e76f-95c0-465f-b059-188802063bac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:26:30 crc kubenswrapper[4945]: I1008 15:26:30.909931 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d0e76f-95c0-465f-b059-188802063bac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.481687 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" event={"ID":"d39d5733-b343-4a33-866b-220a8191eda9","Type":"ContainerStarted","Data":"51c515072c0a9b340af2bc2f83bfef6232d7d9d406314d5604e8d722933b805b"} Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.487224 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdq9c" event={"ID":"f4d0e76f-95c0-465f-b059-188802063bac","Type":"ContainerDied","Data":"4e4532a0418b109469e7f6246fef22925747e7aa8d1fd1aabca69dc1cb5b92f3"} Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.487478 4945 scope.go:117] "RemoveContainer" containerID="33ec83319ffa8baede2ad01eba1043aa5f1ff873d51da17ba0eb2fbd91e6b225" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.487257 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdq9c" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.488493 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" event={"ID":"09a37473-34c8-43ff-a839-3f712f3d0605","Type":"ContainerStarted","Data":"d445fbea88faf2d2608ba4028e505d1273a23e4b44a083c0457a98e776325972"} Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.488684 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.490549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" event={"ID":"fed6f952-f438-4a3b-bd8c-f2832a1c6e8d","Type":"ContainerStarted","Data":"0076d7bd4f1ed8451753e1126d33d00de0cb4765805de88e5c5ef985f85ed150"} Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.491879 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" event={"ID":"658dd3b7-2709-4785-aa43-64f041d46b7a","Type":"ContainerStarted","Data":"83fb73fa6c3e506db424e14c734cfcb60c5e74a6d23834111d0b9b26e77acf4a"} Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.493763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" event={"ID":"18169f3c-d245-4b63-8ea7-d6d3bd70cce4","Type":"ContainerStarted","Data":"7aa30d5f49703395a106a56fb66cbe0af9e7d0b61a51f5361be63b13a528647f"} Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.505547 4945 scope.go:117] "RemoveContainer" containerID="1f97ee03f05bd2e983bb2c289a62311383cd354a37e3fbf29b0f986239e26b5c" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.509747 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" podStartSLOduration=1.815816264 podStartE2EDuration="17.509711248s" podCreationTimestamp="2025-10-08 15:26:14 +0000 UTC" firstStartedPulling="2025-10-08 15:26:15.097754506 +0000 UTC m=+784.451669407" lastFinishedPulling="2025-10-08 15:26:30.79164949 +0000 UTC m=+800.145564391" observedRunningTime="2025-10-08 15:26:31.499518449 +0000 UTC m=+800.853433350" watchObservedRunningTime="2025-10-08 15:26:31.509711248 +0000 UTC m=+800.863626149" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.511846 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.549386 4945 scope.go:117] "RemoveContainer" containerID="b58a18e3c9fa070fb4412123ece15312fc95d9e3a55f44f6aa8f397554ee6f8d" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.558849 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp" podStartSLOduration=2.595476987 podStartE2EDuration="18.558831127s" podCreationTimestamp="2025-10-08 15:26:13 +0000 UTC" firstStartedPulling="2025-10-08 15:26:14.828347182 +0000 UTC m=+784.182262083" lastFinishedPulling="2025-10-08 15:26:30.791701332 +0000 UTC m=+800.145616223" observedRunningTime="2025-10-08 15:26:31.53058018 +0000 UTC m=+800.884495081" watchObservedRunningTime="2025-10-08 15:26:31.558831127 +0000 UTC m=+800.912746028" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.561370 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd" podStartSLOduration=2.57207914 podStartE2EDuration="18.561361144s" podCreationTimestamp="2025-10-08 15:26:13 +0000 UTC" firstStartedPulling="2025-10-08 15:26:14.818273936 +0000 UTC m=+784.172188837" lastFinishedPulling="2025-10-08 15:26:30.80755593 +0000 UTC m=+800.161470841" observedRunningTime="2025-10-08 15:26:31.556219728 +0000 UTC m=+800.910134629" watchObservedRunningTime="2025-10-08 15:26:31.561361144 +0000 UTC m=+800.915276055" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.591850 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-qrx7m" podStartSLOduration=1.716430206 podStartE2EDuration="17.59183629s" podCreationTimestamp="2025-10-08 15:26:14 +0000 UTC" firstStartedPulling="2025-10-08 15:26:14.971135878 +0000 UTC m=+784.325050769" lastFinishedPulling="2025-10-08 15:26:30.846541952 +0000 UTC m=+800.200456853" observedRunningTime="2025-10-08 15:26:31.589526599 +0000 UTC m=+800.943441490" watchObservedRunningTime="2025-10-08 15:26:31.59183629 +0000 UTC m=+800.945751191" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.621913 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-g4w8k" podStartSLOduration=2.528203428 podStartE2EDuration="18.621893255s" podCreationTimestamp="2025-10-08 15:26:13 +0000 UTC" firstStartedPulling="2025-10-08 15:26:14.722699998 +0000 UTC m=+784.076614909" lastFinishedPulling="2025-10-08 15:26:30.816389835 +0000 UTC m=+800.170304736" observedRunningTime="2025-10-08 15:26:31.6134107 +0000 UTC m=+800.967325601" watchObservedRunningTime="2025-10-08 15:26:31.621893255 +0000 UTC m=+800.975808156" Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.645719 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qdq9c"] Oct 08 15:26:31 crc kubenswrapper[4945]: I1008 15:26:31.649397 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qdq9c"] Oct 08 15:26:32 crc kubenswrapper[4945]: I1008 15:26:32.030796 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4d0e76f-95c0-465f-b059-188802063bac" path="/var/lib/kubelet/pods/f4d0e76f-95c0-465f-b059-188802063bac/volumes" Oct 08 15:26:32 crc kubenswrapper[4945]: I1008 15:26:32.499509 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:44 crc kubenswrapper[4945]: I1008 15:26:44.534871 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-sw8w6" Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.183881 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.184510 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.184562 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.185192 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec80442af9d49173b57d5cd8aba7514d1075212c202e9f51e35b871dde889d11"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.185257 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://ec80442af9d49173b57d5cd8aba7514d1075212c202e9f51e35b871dde889d11" gracePeriod=600 Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.594365 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="ec80442af9d49173b57d5cd8aba7514d1075212c202e9f51e35b871dde889d11" exitCode=0 Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.594448 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"ec80442af9d49173b57d5cd8aba7514d1075212c202e9f51e35b871dde889d11"} Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.594644 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"06c6bd45fc6832af457ec9bd8f12ba1218de606e23dd6ce0bb0b843b24d0f361"} Oct 08 15:26:49 crc kubenswrapper[4945]: I1008 15:26:49.594667 4945 scope.go:117] "RemoveContainer" containerID="92e8cf284bb0b21015533deb9be10a938603af77ea1c43c763cb1d3ffcf03af1" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.849022 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb"] Oct 08 15:27:07 crc kubenswrapper[4945]: E1008 15:27:07.849854 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="extract-content" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.849870 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="extract-content" Oct 08 15:27:07 crc kubenswrapper[4945]: E1008 15:27:07.849883 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="extract-utilities" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.849890 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="extract-utilities" Oct 08 15:27:07 crc kubenswrapper[4945]: E1008 15:27:07.849899 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="registry-server" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.849906 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="registry-server" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.850020 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4d0e76f-95c0-465f-b059-188802063bac" containerName="registry-server" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.850943 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.853007 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.858154 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb"] Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.984735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cqg8\" (UniqueName: \"kubernetes.io/projected/9b9fcc55-1a43-4136-aebc-50597faeb89d-kube-api-access-7cqg8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.984832 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:07 crc kubenswrapper[4945]: I1008 15:27:07.984964 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.086162 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.086417 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cqg8\" (UniqueName: \"kubernetes.io/projected/9b9fcc55-1a43-4136-aebc-50597faeb89d-kube-api-access-7cqg8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.086518 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.087005 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.087020 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.107638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cqg8\" (UniqueName: \"kubernetes.io/projected/9b9fcc55-1a43-4136-aebc-50597faeb89d-kube-api-access-7cqg8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.168582 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.356667 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb"] Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.711151 4945 generic.go:334] "Generic (PLEG): container finished" podID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerID="c1cb240d0c8888c98049586928f4ff7fe2c5636c434a6cd3aec352140930d3de" exitCode=0 Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.711246 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" event={"ID":"9b9fcc55-1a43-4136-aebc-50597faeb89d","Type":"ContainerDied","Data":"c1cb240d0c8888c98049586928f4ff7fe2c5636c434a6cd3aec352140930d3de"} Oct 08 15:27:08 crc kubenswrapper[4945]: I1008 15:27:08.711865 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" event={"ID":"9b9fcc55-1a43-4136-aebc-50597faeb89d","Type":"ContainerStarted","Data":"8a69f28dd50cdb6560d4c9dfe1d814ab13741e2766bc62f46c5fafde53e0556f"} Oct 08 15:27:11 crc kubenswrapper[4945]: I1008 15:27:11.731770 4945 generic.go:334] "Generic (PLEG): container finished" podID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerID="2a13788eb3026d02763c8ba198c3346b3b0ce518b516e7cd59c6d8dbc09e2c49" exitCode=0 Oct 08 15:27:11 crc kubenswrapper[4945]: I1008 15:27:11.731917 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" event={"ID":"9b9fcc55-1a43-4136-aebc-50597faeb89d","Type":"ContainerDied","Data":"2a13788eb3026d02763c8ba198c3346b3b0ce518b516e7cd59c6d8dbc09e2c49"} Oct 08 15:27:12 crc kubenswrapper[4945]: I1008 15:27:12.740489 4945 generic.go:334] "Generic (PLEG): container finished" podID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerID="5ae93987bdc017460f38c87cf9816c53284bd261fa2995cbb6d8e5c64346fefa" exitCode=0 Oct 08 15:27:12 crc kubenswrapper[4945]: I1008 15:27:12.740536 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" event={"ID":"9b9fcc55-1a43-4136-aebc-50597faeb89d","Type":"ContainerDied","Data":"5ae93987bdc017460f38c87cf9816c53284bd261fa2995cbb6d8e5c64346fefa"} Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.022635 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.162589 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-util\") pod \"9b9fcc55-1a43-4136-aebc-50597faeb89d\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.162727 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cqg8\" (UniqueName: \"kubernetes.io/projected/9b9fcc55-1a43-4136-aebc-50597faeb89d-kube-api-access-7cqg8\") pod \"9b9fcc55-1a43-4136-aebc-50597faeb89d\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.162760 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-bundle\") pod \"9b9fcc55-1a43-4136-aebc-50597faeb89d\" (UID: \"9b9fcc55-1a43-4136-aebc-50597faeb89d\") " Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.163778 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-bundle" (OuterVolumeSpecName: "bundle") pod "9b9fcc55-1a43-4136-aebc-50597faeb89d" (UID: "9b9fcc55-1a43-4136-aebc-50597faeb89d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.167853 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b9fcc55-1a43-4136-aebc-50597faeb89d-kube-api-access-7cqg8" (OuterVolumeSpecName: "kube-api-access-7cqg8") pod "9b9fcc55-1a43-4136-aebc-50597faeb89d" (UID: "9b9fcc55-1a43-4136-aebc-50597faeb89d"). InnerVolumeSpecName "kube-api-access-7cqg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.264331 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cqg8\" (UniqueName: \"kubernetes.io/projected/9b9fcc55-1a43-4136-aebc-50597faeb89d-kube-api-access-7cqg8\") on node \"crc\" DevicePath \"\"" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.264376 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.342692 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-util" (OuterVolumeSpecName: "util") pod "9b9fcc55-1a43-4136-aebc-50597faeb89d" (UID: "9b9fcc55-1a43-4136-aebc-50597faeb89d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.365511 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b9fcc55-1a43-4136-aebc-50597faeb89d-util\") on node \"crc\" DevicePath \"\"" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.755028 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" event={"ID":"9b9fcc55-1a43-4136-aebc-50597faeb89d","Type":"ContainerDied","Data":"8a69f28dd50cdb6560d4c9dfe1d814ab13741e2766bc62f46c5fafde53e0556f"} Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.755066 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a69f28dd50cdb6560d4c9dfe1d814ab13741e2766bc62f46c5fafde53e0556f" Oct 08 15:27:14 crc kubenswrapper[4945]: I1008 15:27:14.755085 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.197466 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-z75g2"] Oct 08 15:27:19 crc kubenswrapper[4945]: E1008 15:27:19.198155 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerName="util" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.198167 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerName="util" Oct 08 15:27:19 crc kubenswrapper[4945]: E1008 15:27:19.198177 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerName="extract" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.198184 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerName="extract" Oct 08 15:27:19 crc kubenswrapper[4945]: E1008 15:27:19.198198 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerName="pull" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.198204 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerName="pull" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.198300 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b9fcc55-1a43-4136-aebc-50597faeb89d" containerName="extract" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.198663 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.201009 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-d86qd" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.201181 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.201965 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.219057 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-z75g2"] Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.324681 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtdhb\" (UniqueName: \"kubernetes.io/projected/9d7e5f1a-5f1b-4291-8dee-de24611c5ff7-kube-api-access-wtdhb\") pod \"nmstate-operator-858ddd8f98-z75g2\" (UID: \"9d7e5f1a-5f1b-4291-8dee-de24611c5ff7\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.425787 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtdhb\" (UniqueName: \"kubernetes.io/projected/9d7e5f1a-5f1b-4291-8dee-de24611c5ff7-kube-api-access-wtdhb\") pod \"nmstate-operator-858ddd8f98-z75g2\" (UID: \"9d7e5f1a-5f1b-4291-8dee-de24611c5ff7\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.447723 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtdhb\" (UniqueName: \"kubernetes.io/projected/9d7e5f1a-5f1b-4291-8dee-de24611c5ff7-kube-api-access-wtdhb\") pod \"nmstate-operator-858ddd8f98-z75g2\" (UID: \"9d7e5f1a-5f1b-4291-8dee-de24611c5ff7\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.513409 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" Oct 08 15:27:19 crc kubenswrapper[4945]: I1008 15:27:19.912130 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-z75g2"] Oct 08 15:27:20 crc kubenswrapper[4945]: I1008 15:27:20.786692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" event={"ID":"9d7e5f1a-5f1b-4291-8dee-de24611c5ff7","Type":"ContainerStarted","Data":"2a441af6abcdcbe013c9877b870a1f65f58c494c411dbb4bc917cc8877c48a90"} Oct 08 15:27:22 crc kubenswrapper[4945]: I1008 15:27:22.798078 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" event={"ID":"9d7e5f1a-5f1b-4291-8dee-de24611c5ff7","Type":"ContainerStarted","Data":"b9d2b1bef78f0144f41aeca5d7c070496303c470da3f9b0ae0f79ce925005f7e"} Oct 08 15:27:22 crc kubenswrapper[4945]: I1008 15:27:22.818478 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z75g2" podStartSLOduration=1.532439744 podStartE2EDuration="3.818459774s" podCreationTimestamp="2025-10-08 15:27:19 +0000 UTC" firstStartedPulling="2025-10-08 15:27:19.920589784 +0000 UTC m=+849.274504685" lastFinishedPulling="2025-10-08 15:27:22.206609814 +0000 UTC m=+851.560524715" observedRunningTime="2025-10-08 15:27:22.814697373 +0000 UTC m=+852.168612284" watchObservedRunningTime="2025-10-08 15:27:22.818459774 +0000 UTC m=+852.172374695" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.138901 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2675c"] Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.140703 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.157139 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2675c"] Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.320784 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz5q6\" (UniqueName: \"kubernetes.io/projected/ddb9e755-4d6f-42aa-8655-294bc2258708-kube-api-access-xz5q6\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.320909 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-catalog-content\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.320977 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-utilities\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.422152 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz5q6\" (UniqueName: \"kubernetes.io/projected/ddb9e755-4d6f-42aa-8655-294bc2258708-kube-api-access-xz5q6\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.422266 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-catalog-content\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.422328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-utilities\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.422804 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-catalog-content\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.423098 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-utilities\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.440890 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz5q6\" (UniqueName: \"kubernetes.io/projected/ddb9e755-4d6f-42aa-8655-294bc2258708-kube-api-access-xz5q6\") pod \"community-operators-2675c\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.461561 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.673337 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2675c"] Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.820974 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2675c" event={"ID":"ddb9e755-4d6f-42aa-8655-294bc2258708","Type":"ContainerStarted","Data":"ba87160a4e5a3ef2460457f4424143b822a1a53af1161bf0c5a9663224b1591d"} Oct 08 15:27:26 crc kubenswrapper[4945]: I1008 15:27:26.821018 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2675c" event={"ID":"ddb9e755-4d6f-42aa-8655-294bc2258708","Type":"ContainerStarted","Data":"59e3a34f4a4dbb725b57679e07b6aa41c293637772bcbd82a143b9e0ef268944"} Oct 08 15:27:27 crc kubenswrapper[4945]: I1008 15:27:27.829133 4945 generic.go:334] "Generic (PLEG): container finished" podID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerID="ba87160a4e5a3ef2460457f4424143b822a1a53af1161bf0c5a9663224b1591d" exitCode=0 Oct 08 15:27:27 crc kubenswrapper[4945]: I1008 15:27:27.829242 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2675c" event={"ID":"ddb9e755-4d6f-42aa-8655-294bc2258708","Type":"ContainerDied","Data":"ba87160a4e5a3ef2460457f4424143b822a1a53af1161bf0c5a9663224b1591d"} Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.063015 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.064222 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" Oct 08 15:27:29 crc kubenswrapper[4945]: W1008 15:27:29.071027 4945 reflector.go:561] object-"openshift-nmstate"/"nmstate-handler-dockercfg-jbc5h": failed to list *v1.Secret: secrets "nmstate-handler-dockercfg-jbc5h" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Oct 08 15:27:29 crc kubenswrapper[4945]: E1008 15:27:29.071074 4945 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"nmstate-handler-dockercfg-jbc5h\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nmstate-handler-dockercfg-jbc5h\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.095139 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.098833 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.109799 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.113913 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.119007 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.123077 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6q4xl"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.124091 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.215162 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.216088 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.218847 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.219160 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.219220 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-zvrgm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.226280 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.259900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwzf5\" (UniqueName: \"kubernetes.io/projected/0afd8411-7a96-44b9-b1af-82a05419fa24-kube-api-access-nwzf5\") pod \"nmstate-metrics-fdff9cb8d-kfqmx\" (UID: \"0afd8411-7a96-44b9-b1af-82a05419fa24\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.259942 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c84bb408-d873-4b03-a07b-dfb913fdc64b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-59hrm\" (UID: \"c84bb408-d873-4b03-a07b-dfb913fdc64b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.259960 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-dbus-socket\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.259997 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h85ml\" (UniqueName: \"kubernetes.io/projected/835b6fb9-28f9-4c64-8049-05dd404be8cc-kube-api-access-h85ml\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.260018 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tcs7\" (UniqueName: \"kubernetes.io/projected/c84bb408-d873-4b03-a07b-dfb913fdc64b-kube-api-access-7tcs7\") pod \"nmstate-webhook-6cdbc54649-59hrm\" (UID: \"c84bb408-d873-4b03-a07b-dfb913fdc64b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.260084 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-nmstate-lock\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.260145 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-ovs-socket\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9183c-cff3-4bed-82e0-41a654969322-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2kvr\" (UniqueName: \"kubernetes.io/projected/d4d9183c-cff3-4bed-82e0-41a654969322-kube-api-access-x2kvr\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361593 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-nmstate-lock\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361661 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d4d9183c-cff3-4bed-82e0-41a654969322-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-ovs-socket\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-nmstate-lock\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361750 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-ovs-socket\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361842 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwzf5\" (UniqueName: \"kubernetes.io/projected/0afd8411-7a96-44b9-b1af-82a05419fa24-kube-api-access-nwzf5\") pod \"nmstate-metrics-fdff9cb8d-kfqmx\" (UID: \"0afd8411-7a96-44b9-b1af-82a05419fa24\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c84bb408-d873-4b03-a07b-dfb913fdc64b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-59hrm\" (UID: \"c84bb408-d873-4b03-a07b-dfb913fdc64b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.361987 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-dbus-socket\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.362020 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h85ml\" (UniqueName: \"kubernetes.io/projected/835b6fb9-28f9-4c64-8049-05dd404be8cc-kube-api-access-h85ml\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.362051 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tcs7\" (UniqueName: \"kubernetes.io/projected/c84bb408-d873-4b03-a07b-dfb913fdc64b-kube-api-access-7tcs7\") pod \"nmstate-webhook-6cdbc54649-59hrm\" (UID: \"c84bb408-d873-4b03-a07b-dfb913fdc64b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.362398 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/835b6fb9-28f9-4c64-8049-05dd404be8cc-dbus-socket\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.367973 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c84bb408-d873-4b03-a07b-dfb913fdc64b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-59hrm\" (UID: \"c84bb408-d873-4b03-a07b-dfb913fdc64b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.380853 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tcs7\" (UniqueName: \"kubernetes.io/projected/c84bb408-d873-4b03-a07b-dfb913fdc64b-kube-api-access-7tcs7\") pod \"nmstate-webhook-6cdbc54649-59hrm\" (UID: \"c84bb408-d873-4b03-a07b-dfb913fdc64b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.387678 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h85ml\" (UniqueName: \"kubernetes.io/projected/835b6fb9-28f9-4c64-8049-05dd404be8cc-kube-api-access-h85ml\") pod \"nmstate-handler-6q4xl\" (UID: \"835b6fb9-28f9-4c64-8049-05dd404be8cc\") " pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.388813 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwzf5\" (UniqueName: \"kubernetes.io/projected/0afd8411-7a96-44b9-b1af-82a05419fa24-kube-api-access-nwzf5\") pod \"nmstate-metrics-fdff9cb8d-kfqmx\" (UID: \"0afd8411-7a96-44b9-b1af-82a05419fa24\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.406664 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f59c67f96-znllz"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.407458 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.417283 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f59c67f96-znllz"] Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.463667 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9183c-cff3-4bed-82e0-41a654969322-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.463702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2kvr\" (UniqueName: \"kubernetes.io/projected/d4d9183c-cff3-4bed-82e0-41a654969322-kube-api-access-x2kvr\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.463730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d4d9183c-cff3-4bed-82e0-41a654969322-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: E1008 15:27:29.463944 4945 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 08 15:27:29 crc kubenswrapper[4945]: E1008 15:27:29.464047 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4d9183c-cff3-4bed-82e0-41a654969322-plugin-serving-cert podName:d4d9183c-cff3-4bed-82e0-41a654969322 nodeName:}" failed. No retries permitted until 2025-10-08 15:27:29.964032685 +0000 UTC m=+859.317947586 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/d4d9183c-cff3-4bed-82e0-41a654969322-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-7zpln" (UID: "d4d9183c-cff3-4bed-82e0-41a654969322") : secret "plugin-serving-cert" not found Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.464531 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d4d9183c-cff3-4bed-82e0-41a654969322-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.477834 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2kvr\" (UniqueName: \"kubernetes.io/projected/d4d9183c-cff3-4bed-82e0-41a654969322-kube-api-access-x2kvr\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.564954 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-oauth-config\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.564996 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-service-ca\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.565193 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-serving-cert\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.565285 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-config\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.565417 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-oauth-serving-cert\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.565455 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm9n9\" (UniqueName: \"kubernetes.io/projected/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-kube-api-access-sm9n9\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.565487 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-trusted-ca-bundle\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.666705 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-oauth-serving-cert\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.667990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm9n9\" (UniqueName: \"kubernetes.io/projected/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-kube-api-access-sm9n9\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.667747 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-oauth-serving-cert\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.668641 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-trusted-ca-bundle\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.668806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-oauth-config\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.668964 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-service-ca\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.669648 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-serving-cert\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.669789 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-config\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.669838 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-service-ca\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.670580 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-config\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.672352 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-trusted-ca-bundle\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.674171 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-oauth-config\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.674722 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-console-serving-cert\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.685206 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm9n9\" (UniqueName: \"kubernetes.io/projected/ae11f2fc-b6f8-4625-9dd6-264e39639e2b-kube-api-access-sm9n9\") pod \"console-f59c67f96-znllz\" (UID: \"ae11f2fc-b6f8-4625-9dd6-264e39639e2b\") " pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.752032 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.861265 4945 generic.go:334] "Generic (PLEG): container finished" podID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerID="11098c352fa4e1f0f70fa8130eaff4b86839ec1ba7e36702b8b3b2b4370bc530" exitCode=0 Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.861655 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2675c" event={"ID":"ddb9e755-4d6f-42aa-8655-294bc2258708","Type":"ContainerDied","Data":"11098c352fa4e1f0f70fa8130eaff4b86839ec1ba7e36702b8b3b2b4370bc530"} Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.975018 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9183c-cff3-4bed-82e0-41a654969322-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:29 crc kubenswrapper[4945]: I1008 15:27:29.981829 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9183c-cff3-4bed-82e0-41a654969322-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-7zpln\" (UID: \"d4d9183c-cff3-4bed-82e0-41a654969322\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.131581 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.189704 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f59c67f96-znllz"] Oct 08 15:27:30 crc kubenswrapper[4945]: W1008 15:27:30.200036 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae11f2fc_b6f8_4625_9dd6_264e39639e2b.slice/crio-8c31015da8983127d8293a0e0e6768bb44e3d8ef5a72692e2b87b8fb799a1228 WatchSource:0}: Error finding container 8c31015da8983127d8293a0e0e6768bb44e3d8ef5a72692e2b87b8fb799a1228: Status 404 returned error can't find the container with id 8c31015da8983127d8293a0e0e6768bb44e3d8ef5a72692e2b87b8fb799a1228 Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.352484 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln"] Oct 08 15:27:30 crc kubenswrapper[4945]: W1008 15:27:30.354800 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4d9183c_cff3_4bed_82e0_41a654969322.slice/crio-d4921bf97a33153de49a4749e1d50aeba1473c980d630bd314d820ca728853a2 WatchSource:0}: Error finding container d4921bf97a33153de49a4749e1d50aeba1473c980d630bd314d820ca728853a2: Status 404 returned error can't find the container with id d4921bf97a33153de49a4749e1d50aeba1473c980d630bd314d820ca728853a2 Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.429854 4945 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.430423 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.441293 4945 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-nmstate/nmstate-handler-6q4xl" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.441367 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.540453 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jbc5h" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.548153 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.862999 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm"] Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.868177 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6q4xl" event={"ID":"835b6fb9-28f9-4c64-8049-05dd404be8cc","Type":"ContainerStarted","Data":"dc1ff62b6669cb176009bf2d796ab720896de97e2a2b93431f4a7a1c1252a31f"} Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.869341 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f59c67f96-znllz" event={"ID":"ae11f2fc-b6f8-4625-9dd6-264e39639e2b","Type":"ContainerStarted","Data":"cbcc4d5eea70cafda991c206328dd44ebab6bd65f73e292c308e33253303b089"} Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.869370 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f59c67f96-znllz" event={"ID":"ae11f2fc-b6f8-4625-9dd6-264e39639e2b","Type":"ContainerStarted","Data":"8c31015da8983127d8293a0e0e6768bb44e3d8ef5a72692e2b87b8fb799a1228"} Oct 08 15:27:30 crc kubenswrapper[4945]: W1008 15:27:30.869741 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc84bb408_d873_4b03_a07b_dfb913fdc64b.slice/crio-022bf13dd7c45f2b25b4ead37b9b02673da7d1568bea966e8592b0f9dd8bc29e WatchSource:0}: Error finding container 022bf13dd7c45f2b25b4ead37b9b02673da7d1568bea966e8592b0f9dd8bc29e: Status 404 returned error can't find the container with id 022bf13dd7c45f2b25b4ead37b9b02673da7d1568bea966e8592b0f9dd8bc29e Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.871611 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2675c" event={"ID":"ddb9e755-4d6f-42aa-8655-294bc2258708","Type":"ContainerStarted","Data":"ce581c1425045fd61a7580f9e40f464d9327f1bcfcb05251c0a8fb7c7bfa49f1"} Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.874488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" event={"ID":"d4d9183c-cff3-4bed-82e0-41a654969322","Type":"ContainerStarted","Data":"d4921bf97a33153de49a4749e1d50aeba1473c980d630bd314d820ca728853a2"} Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.889711 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f59c67f96-znllz" podStartSLOduration=1.889693543 podStartE2EDuration="1.889693543s" podCreationTimestamp="2025-10-08 15:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:27:30.88477222 +0000 UTC m=+860.238687121" watchObservedRunningTime="2025-10-08 15:27:30.889693543 +0000 UTC m=+860.243608444" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.904493 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2675c" podStartSLOduration=2.132899641 podStartE2EDuration="4.904478621s" podCreationTimestamp="2025-10-08 15:27:26 +0000 UTC" firstStartedPulling="2025-10-08 15:27:27.830911018 +0000 UTC m=+857.184825949" lastFinishedPulling="2025-10-08 15:27:30.602490028 +0000 UTC m=+859.956404929" observedRunningTime="2025-10-08 15:27:30.898967902 +0000 UTC m=+860.252882803" watchObservedRunningTime="2025-10-08 15:27:30.904478621 +0000 UTC m=+860.258393522" Oct 08 15:27:30 crc kubenswrapper[4945]: I1008 15:27:30.978720 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx"] Oct 08 15:27:30 crc kubenswrapper[4945]: W1008 15:27:30.981309 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0afd8411_7a96_44b9_b1af_82a05419fa24.slice/crio-4d7f2e49f1a459732272f8f53fe596478d4e477853967769803dcfd7a84adfdf WatchSource:0}: Error finding container 4d7f2e49f1a459732272f8f53fe596478d4e477853967769803dcfd7a84adfdf: Status 404 returned error can't find the container with id 4d7f2e49f1a459732272f8f53fe596478d4e477853967769803dcfd7a84adfdf Oct 08 15:27:31 crc kubenswrapper[4945]: I1008 15:27:31.886474 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" event={"ID":"0afd8411-7a96-44b9-b1af-82a05419fa24","Type":"ContainerStarted","Data":"4d7f2e49f1a459732272f8f53fe596478d4e477853967769803dcfd7a84adfdf"} Oct 08 15:27:31 crc kubenswrapper[4945]: I1008 15:27:31.888647 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" event={"ID":"c84bb408-d873-4b03-a07b-dfb913fdc64b","Type":"ContainerStarted","Data":"022bf13dd7c45f2b25b4ead37b9b02673da7d1568bea966e8592b0f9dd8bc29e"} Oct 08 15:27:33 crc kubenswrapper[4945]: I1008 15:27:33.898396 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" event={"ID":"0afd8411-7a96-44b9-b1af-82a05419fa24","Type":"ContainerStarted","Data":"c1c25806489fbdc924f6bbd35133d35a10e5650687ffa3500a93691a6530fbc2"} Oct 08 15:27:33 crc kubenswrapper[4945]: I1008 15:27:33.899521 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" event={"ID":"d4d9183c-cff3-4bed-82e0-41a654969322","Type":"ContainerStarted","Data":"25a07f29b3abd49e91cc05f262c5d0003c0e38aad8e053b9c5e940af4abbde5e"} Oct 08 15:27:33 crc kubenswrapper[4945]: I1008 15:27:33.901463 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" event={"ID":"c84bb408-d873-4b03-a07b-dfb913fdc64b","Type":"ContainerStarted","Data":"2cab7ffbb0ddadb473aa95f047bcf968a26cfb2de873651678868190e8447bef"} Oct 08 15:27:33 crc kubenswrapper[4945]: I1008 15:27:33.901790 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:33 crc kubenswrapper[4945]: I1008 15:27:33.915381 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7zpln" podStartSLOduration=1.670254583 podStartE2EDuration="4.915358815s" podCreationTimestamp="2025-10-08 15:27:29 +0000 UTC" firstStartedPulling="2025-10-08 15:27:30.357305964 +0000 UTC m=+859.711220865" lastFinishedPulling="2025-10-08 15:27:33.602410156 +0000 UTC m=+862.956325097" observedRunningTime="2025-10-08 15:27:33.913344691 +0000 UTC m=+863.267259592" watchObservedRunningTime="2025-10-08 15:27:33.915358815 +0000 UTC m=+863.269273716" Oct 08 15:27:34 crc kubenswrapper[4945]: I1008 15:27:34.909155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6q4xl" event={"ID":"835b6fb9-28f9-4c64-8049-05dd404be8cc","Type":"ContainerStarted","Data":"1370f1b0f417a5e9a70ff3f1de032b0498390b9284e727e517942b3f09c041c2"} Oct 08 15:27:34 crc kubenswrapper[4945]: I1008 15:27:34.909588 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:34 crc kubenswrapper[4945]: I1008 15:27:34.924818 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6q4xl" podStartSLOduration=2.770987448 podStartE2EDuration="5.924799523s" podCreationTimestamp="2025-10-08 15:27:29 +0000 UTC" firstStartedPulling="2025-10-08 15:27:30.468875648 +0000 UTC m=+859.822790559" lastFinishedPulling="2025-10-08 15:27:33.622687723 +0000 UTC m=+862.976602634" observedRunningTime="2025-10-08 15:27:34.923279942 +0000 UTC m=+864.277194853" watchObservedRunningTime="2025-10-08 15:27:34.924799523 +0000 UTC m=+864.278714434" Oct 08 15:27:34 crc kubenswrapper[4945]: I1008 15:27:34.928750 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" podStartSLOduration=3.173794639 podStartE2EDuration="5.928730019s" podCreationTimestamp="2025-10-08 15:27:29 +0000 UTC" firstStartedPulling="2025-10-08 15:27:30.871300408 +0000 UTC m=+860.225215309" lastFinishedPulling="2025-10-08 15:27:33.626235778 +0000 UTC m=+862.980150689" observedRunningTime="2025-10-08 15:27:33.929578498 +0000 UTC m=+863.283493399" watchObservedRunningTime="2025-10-08 15:27:34.928730019 +0000 UTC m=+864.282644930" Oct 08 15:27:36 crc kubenswrapper[4945]: I1008 15:27:36.462661 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:36 crc kubenswrapper[4945]: I1008 15:27:36.463416 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:36 crc kubenswrapper[4945]: I1008 15:27:36.533159 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:36 crc kubenswrapper[4945]: I1008 15:27:36.991203 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:37 crc kubenswrapper[4945]: I1008 15:27:37.938007 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" event={"ID":"0afd8411-7a96-44b9-b1af-82a05419fa24","Type":"ContainerStarted","Data":"7787a4f7afc3704bbbaf8e4bf93a5479f0da63a3832b0b2fe6b4b7f522877d57"} Oct 08 15:27:38 crc kubenswrapper[4945]: I1008 15:27:38.923274 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-kfqmx" podStartSLOduration=3.274539723 podStartE2EDuration="9.923247897s" podCreationTimestamp="2025-10-08 15:27:29 +0000 UTC" firstStartedPulling="2025-10-08 15:27:30.983184661 +0000 UTC m=+860.337099562" lastFinishedPulling="2025-10-08 15:27:37.631892815 +0000 UTC m=+866.985807736" observedRunningTime="2025-10-08 15:27:37.960891316 +0000 UTC m=+867.314806257" watchObservedRunningTime="2025-10-08 15:27:38.923247897 +0000 UTC m=+868.277162808" Oct 08 15:27:38 crc kubenswrapper[4945]: I1008 15:27:38.927532 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2675c"] Oct 08 15:27:39 crc kubenswrapper[4945]: I1008 15:27:39.753023 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:39 crc kubenswrapper[4945]: I1008 15:27:39.753185 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:39 crc kubenswrapper[4945]: I1008 15:27:39.760402 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:39 crc kubenswrapper[4945]: I1008 15:27:39.953396 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2675c" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="registry-server" containerID="cri-o://ce581c1425045fd61a7580f9e40f464d9327f1bcfcb05251c0a8fb7c7bfa49f1" gracePeriod=2 Oct 08 15:27:39 crc kubenswrapper[4945]: I1008 15:27:39.960379 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f59c67f96-znllz" Oct 08 15:27:40 crc kubenswrapper[4945]: I1008 15:27:40.020473 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kr5ql"] Oct 08 15:27:40 crc kubenswrapper[4945]: I1008 15:27:40.465759 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6q4xl" Oct 08 15:27:41 crc kubenswrapper[4945]: I1008 15:27:41.968911 4945 generic.go:334] "Generic (PLEG): container finished" podID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerID="ce581c1425045fd61a7580f9e40f464d9327f1bcfcb05251c0a8fb7c7bfa49f1" exitCode=0 Oct 08 15:27:41 crc kubenswrapper[4945]: I1008 15:27:41.969044 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2675c" event={"ID":"ddb9e755-4d6f-42aa-8655-294bc2258708","Type":"ContainerDied","Data":"ce581c1425045fd61a7580f9e40f464d9327f1bcfcb05251c0a8fb7c7bfa49f1"} Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.792840 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.966574 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-catalog-content\") pod \"ddb9e755-4d6f-42aa-8655-294bc2258708\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.966689 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz5q6\" (UniqueName: \"kubernetes.io/projected/ddb9e755-4d6f-42aa-8655-294bc2258708-kube-api-access-xz5q6\") pod \"ddb9e755-4d6f-42aa-8655-294bc2258708\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.966733 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-utilities\") pod \"ddb9e755-4d6f-42aa-8655-294bc2258708\" (UID: \"ddb9e755-4d6f-42aa-8655-294bc2258708\") " Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.967986 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-utilities" (OuterVolumeSpecName: "utilities") pod "ddb9e755-4d6f-42aa-8655-294bc2258708" (UID: "ddb9e755-4d6f-42aa-8655-294bc2258708"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.977429 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddb9e755-4d6f-42aa-8655-294bc2258708-kube-api-access-xz5q6" (OuterVolumeSpecName: "kube-api-access-xz5q6") pod "ddb9e755-4d6f-42aa-8655-294bc2258708" (UID: "ddb9e755-4d6f-42aa-8655-294bc2258708"). InnerVolumeSpecName "kube-api-access-xz5q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.980283 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2675c" event={"ID":"ddb9e755-4d6f-42aa-8655-294bc2258708","Type":"ContainerDied","Data":"59e3a34f4a4dbb725b57679e07b6aa41c293637772bcbd82a143b9e0ef268944"} Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.980321 4945 scope.go:117] "RemoveContainer" containerID="ce581c1425045fd61a7580f9e40f464d9327f1bcfcb05251c0a8fb7c7bfa49f1" Oct 08 15:27:42 crc kubenswrapper[4945]: I1008 15:27:42.980393 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2675c" Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.012124 4945 scope.go:117] "RemoveContainer" containerID="11098c352fa4e1f0f70fa8130eaff4b86839ec1ba7e36702b8b3b2b4370bc530" Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.027663 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddb9e755-4d6f-42aa-8655-294bc2258708" (UID: "ddb9e755-4d6f-42aa-8655-294bc2258708"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.034658 4945 scope.go:117] "RemoveContainer" containerID="ba87160a4e5a3ef2460457f4424143b822a1a53af1161bf0c5a9663224b1591d" Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.068010 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz5q6\" (UniqueName: \"kubernetes.io/projected/ddb9e755-4d6f-42aa-8655-294bc2258708-kube-api-access-xz5q6\") on node \"crc\" DevicePath \"\"" Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.068074 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.068086 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddb9e755-4d6f-42aa-8655-294bc2258708-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.326084 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2675c"] Oct 08 15:27:43 crc kubenswrapper[4945]: I1008 15:27:43.333659 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2675c"] Oct 08 15:27:44 crc kubenswrapper[4945]: I1008 15:27:44.039240 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" path="/var/lib/kubelet/pods/ddb9e755-4d6f-42aa-8655-294bc2258708/volumes" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.745807 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wc8r4"] Oct 08 15:27:46 crc kubenswrapper[4945]: E1008 15:27:46.746535 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="extract-utilities" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.746558 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="extract-utilities" Oct 08 15:27:46 crc kubenswrapper[4945]: E1008 15:27:46.746579 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="registry-server" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.746592 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="registry-server" Oct 08 15:27:46 crc kubenswrapper[4945]: E1008 15:27:46.746606 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="extract-content" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.746620 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="extract-content" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.746810 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddb9e755-4d6f-42aa-8655-294bc2258708" containerName="registry-server" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.748333 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.768059 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wc8r4"] Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.922157 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-catalog-content\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.922281 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ckxd\" (UniqueName: \"kubernetes.io/projected/5af5d89c-9737-4c74-b763-a850596fc7e0-kube-api-access-6ckxd\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:46 crc kubenswrapper[4945]: I1008 15:27:46.922359 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-utilities\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.023792 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ckxd\" (UniqueName: \"kubernetes.io/projected/5af5d89c-9737-4c74-b763-a850596fc7e0-kube-api-access-6ckxd\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.023848 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-utilities\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.023891 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-catalog-content\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.024420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-catalog-content\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.024563 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-utilities\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.048178 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ckxd\" (UniqueName: \"kubernetes.io/projected/5af5d89c-9737-4c74-b763-a850596fc7e0-kube-api-access-6ckxd\") pod \"redhat-marketplace-wc8r4\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.087878 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:47 crc kubenswrapper[4945]: I1008 15:27:47.509495 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wc8r4"] Oct 08 15:27:47 crc kubenswrapper[4945]: W1008 15:27:47.517702 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5af5d89c_9737_4c74_b763_a850596fc7e0.slice/crio-d796e13acfc13e80a6479d9a8d4d2de4141cf1caf7bd636213c20241f1890c14 WatchSource:0}: Error finding container d796e13acfc13e80a6479d9a8d4d2de4141cf1caf7bd636213c20241f1890c14: Status 404 returned error can't find the container with id d796e13acfc13e80a6479d9a8d4d2de4141cf1caf7bd636213c20241f1890c14 Oct 08 15:27:48 crc kubenswrapper[4945]: I1008 15:27:48.015398 4945 generic.go:334] "Generic (PLEG): container finished" podID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerID="a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec" exitCode=0 Oct 08 15:27:48 crc kubenswrapper[4945]: I1008 15:27:48.015509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wc8r4" event={"ID":"5af5d89c-9737-4c74-b763-a850596fc7e0","Type":"ContainerDied","Data":"a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec"} Oct 08 15:27:48 crc kubenswrapper[4945]: I1008 15:27:48.015790 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wc8r4" event={"ID":"5af5d89c-9737-4c74-b763-a850596fc7e0","Type":"ContainerStarted","Data":"d796e13acfc13e80a6479d9a8d4d2de4141cf1caf7bd636213c20241f1890c14"} Oct 08 15:27:50 crc kubenswrapper[4945]: I1008 15:27:50.440330 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-59hrm" Oct 08 15:27:51 crc kubenswrapper[4945]: I1008 15:27:51.039853 4945 generic.go:334] "Generic (PLEG): container finished" podID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerID="fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c" exitCode=0 Oct 08 15:27:51 crc kubenswrapper[4945]: I1008 15:27:51.039933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wc8r4" event={"ID":"5af5d89c-9737-4c74-b763-a850596fc7e0","Type":"ContainerDied","Data":"fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c"} Oct 08 15:27:52 crc kubenswrapper[4945]: I1008 15:27:52.049533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wc8r4" event={"ID":"5af5d89c-9737-4c74-b763-a850596fc7e0","Type":"ContainerStarted","Data":"3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e"} Oct 08 15:27:52 crc kubenswrapper[4945]: I1008 15:27:52.074866 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wc8r4" podStartSLOduration=4.261522077 podStartE2EDuration="6.074845507s" podCreationTimestamp="2025-10-08 15:27:46 +0000 UTC" firstStartedPulling="2025-10-08 15:27:50.02998601 +0000 UTC m=+879.383900951" lastFinishedPulling="2025-10-08 15:27:51.84330948 +0000 UTC m=+881.197224381" observedRunningTime="2025-10-08 15:27:52.073296294 +0000 UTC m=+881.427211195" watchObservedRunningTime="2025-10-08 15:27:52.074845507 +0000 UTC m=+881.428760418" Oct 08 15:27:57 crc kubenswrapper[4945]: I1008 15:27:57.088424 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:57 crc kubenswrapper[4945]: I1008 15:27:57.089000 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:57 crc kubenswrapper[4945]: I1008 15:27:57.136580 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:58 crc kubenswrapper[4945]: I1008 15:27:58.147343 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:27:59 crc kubenswrapper[4945]: I1008 15:27:59.538087 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wc8r4"] Oct 08 15:28:00 crc kubenswrapper[4945]: I1008 15:28:00.098770 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wc8r4" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="registry-server" containerID="cri-o://3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e" gracePeriod=2 Oct 08 15:28:00 crc kubenswrapper[4945]: I1008 15:28:00.985740 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.108354 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-utilities\") pod \"5af5d89c-9737-4c74-b763-a850596fc7e0\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.108433 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-catalog-content\") pod \"5af5d89c-9737-4c74-b763-a850596fc7e0\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.108497 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ckxd\" (UniqueName: \"kubernetes.io/projected/5af5d89c-9737-4c74-b763-a850596fc7e0-kube-api-access-6ckxd\") pod \"5af5d89c-9737-4c74-b763-a850596fc7e0\" (UID: \"5af5d89c-9737-4c74-b763-a850596fc7e0\") " Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.110848 4945 generic.go:334] "Generic (PLEG): container finished" podID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerID="3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e" exitCode=0 Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.110892 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wc8r4" event={"ID":"5af5d89c-9737-4c74-b763-a850596fc7e0","Type":"ContainerDied","Data":"3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e"} Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.110899 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wc8r4" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.110923 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wc8r4" event={"ID":"5af5d89c-9737-4c74-b763-a850596fc7e0","Type":"ContainerDied","Data":"d796e13acfc13e80a6479d9a8d4d2de4141cf1caf7bd636213c20241f1890c14"} Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.110943 4945 scope.go:117] "RemoveContainer" containerID="3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.111001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-utilities" (OuterVolumeSpecName: "utilities") pod "5af5d89c-9737-4c74-b763-a850596fc7e0" (UID: "5af5d89c-9737-4c74-b763-a850596fc7e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.118162 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af5d89c-9737-4c74-b763-a850596fc7e0-kube-api-access-6ckxd" (OuterVolumeSpecName: "kube-api-access-6ckxd") pod "5af5d89c-9737-4c74-b763-a850596fc7e0" (UID: "5af5d89c-9737-4c74-b763-a850596fc7e0"). InnerVolumeSpecName "kube-api-access-6ckxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.123941 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5af5d89c-9737-4c74-b763-a850596fc7e0" (UID: "5af5d89c-9737-4c74-b763-a850596fc7e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.142632 4945 scope.go:117] "RemoveContainer" containerID="fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.158206 4945 scope.go:117] "RemoveContainer" containerID="a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.173591 4945 scope.go:117] "RemoveContainer" containerID="3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e" Oct 08 15:28:01 crc kubenswrapper[4945]: E1008 15:28:01.174165 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e\": container with ID starting with 3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e not found: ID does not exist" containerID="3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.174218 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e"} err="failed to get container status \"3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e\": rpc error: code = NotFound desc = could not find container \"3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e\": container with ID starting with 3376af396caa4df6c4206bc1532f73abd376745741f075b057099b1e2dd92f1e not found: ID does not exist" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.174259 4945 scope.go:117] "RemoveContainer" containerID="fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c" Oct 08 15:28:01 crc kubenswrapper[4945]: E1008 15:28:01.174653 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c\": container with ID starting with fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c not found: ID does not exist" containerID="fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.174677 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c"} err="failed to get container status \"fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c\": rpc error: code = NotFound desc = could not find container \"fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c\": container with ID starting with fcc435129a865e5a9d4e0b56f90862be307f81aebbede9d82881141efad55d9c not found: ID does not exist" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.174692 4945 scope.go:117] "RemoveContainer" containerID="a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec" Oct 08 15:28:01 crc kubenswrapper[4945]: E1008 15:28:01.174924 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec\": container with ID starting with a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec not found: ID does not exist" containerID="a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.174961 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec"} err="failed to get container status \"a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec\": rpc error: code = NotFound desc = could not find container \"a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec\": container with ID starting with a748d0d976caf6c616e48d5f4d3b606b4725fc26423634dc189c50d766258bec not found: ID does not exist" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.211665 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.211700 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ckxd\" (UniqueName: \"kubernetes.io/projected/5af5d89c-9737-4c74-b763-a850596fc7e0-kube-api-access-6ckxd\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.211714 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af5d89c-9737-4c74-b763-a850596fc7e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.464785 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wc8r4"] Oct 08 15:28:01 crc kubenswrapper[4945]: I1008 15:28:01.469761 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wc8r4"] Oct 08 15:28:02 crc kubenswrapper[4945]: I1008 15:28:02.124534 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" path="/var/lib/kubelet/pods/5af5d89c-9737-4c74-b763-a850596fc7e0/volumes" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.336243 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p9tgz"] Oct 08 15:28:03 crc kubenswrapper[4945]: E1008 15:28:03.338614 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="extract-utilities" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.338633 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="extract-utilities" Oct 08 15:28:03 crc kubenswrapper[4945]: E1008 15:28:03.338659 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="extract-content" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.338672 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="extract-content" Oct 08 15:28:03 crc kubenswrapper[4945]: E1008 15:28:03.338693 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="registry-server" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.338701 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="registry-server" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.338804 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af5d89c-9737-4c74-b763-a850596fc7e0" containerName="registry-server" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.340007 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.341750 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9tgz"] Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.444500 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-catalog-content\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.444635 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-utilities\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.444738 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw8zg\" (UniqueName: \"kubernetes.io/projected/207715ea-7f62-46c7-8351-f56bd1750ecd-kube-api-access-pw8zg\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.545849 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-catalog-content\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.546207 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-utilities\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.546245 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw8zg\" (UniqueName: \"kubernetes.io/projected/207715ea-7f62-46c7-8351-f56bd1750ecd-kube-api-access-pw8zg\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.546338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-catalog-content\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.546602 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-utilities\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.574252 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw8zg\" (UniqueName: \"kubernetes.io/projected/207715ea-7f62-46c7-8351-f56bd1750ecd-kube-api-access-pw8zg\") pod \"certified-operators-p9tgz\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.669244 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:03 crc kubenswrapper[4945]: I1008 15:28:03.929373 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9tgz"] Oct 08 15:28:04 crc kubenswrapper[4945]: I1008 15:28:04.158415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9tgz" event={"ID":"207715ea-7f62-46c7-8351-f56bd1750ecd","Type":"ContainerStarted","Data":"d28d5de9c4acd4e05d70e045fdaf5f10b98a8152635df13f051d610d10585b0c"} Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.110575 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-kr5ql" podUID="6df6697f-bec8-4b87-a0ef-7e71c9a0b360" containerName="console" containerID="cri-o://ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e" gracePeriod=15 Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.173577 4945 generic.go:334] "Generic (PLEG): container finished" podID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerID="88b8b91c9a7c33a2e7391012d95fb7243be148c6f92440a2936557d0da2d31bd" exitCode=0 Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.173633 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9tgz" event={"ID":"207715ea-7f62-46c7-8351-f56bd1750ecd","Type":"ContainerDied","Data":"88b8b91c9a7c33a2e7391012d95fb7243be148c6f92440a2936557d0da2d31bd"} Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.566171 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kr5ql_6df6697f-bec8-4b87-a0ef-7e71c9a0b360/console/0.log" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.566446 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679006 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-oauth-config\") pod \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679085 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-service-ca\") pod \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679186 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzmhx\" (UniqueName: \"kubernetes.io/projected/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-kube-api-access-nzmhx\") pod \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679231 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-config\") pod \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679284 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-oauth-serving-cert\") pod \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679311 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-trusted-ca-bundle\") pod \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679435 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-serving-cert\") pod \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\" (UID: \"6df6697f-bec8-4b87-a0ef-7e71c9a0b360\") " Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679943 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-config" (OuterVolumeSpecName: "console-config") pod "6df6697f-bec8-4b87-a0ef-7e71c9a0b360" (UID: "6df6697f-bec8-4b87-a0ef-7e71c9a0b360"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.679985 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-service-ca" (OuterVolumeSpecName: "service-ca") pod "6df6697f-bec8-4b87-a0ef-7e71c9a0b360" (UID: "6df6697f-bec8-4b87-a0ef-7e71c9a0b360"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.680047 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6df6697f-bec8-4b87-a0ef-7e71c9a0b360" (UID: "6df6697f-bec8-4b87-a0ef-7e71c9a0b360"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.680074 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6df6697f-bec8-4b87-a0ef-7e71c9a0b360" (UID: "6df6697f-bec8-4b87-a0ef-7e71c9a0b360"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.693460 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6df6697f-bec8-4b87-a0ef-7e71c9a0b360" (UID: "6df6697f-bec8-4b87-a0ef-7e71c9a0b360"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.693674 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6df6697f-bec8-4b87-a0ef-7e71c9a0b360" (UID: "6df6697f-bec8-4b87-a0ef-7e71c9a0b360"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.695586 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-kube-api-access-nzmhx" (OuterVolumeSpecName: "kube-api-access-nzmhx") pod "6df6697f-bec8-4b87-a0ef-7e71c9a0b360" (UID: "6df6697f-bec8-4b87-a0ef-7e71c9a0b360"). InnerVolumeSpecName "kube-api-access-nzmhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.780676 4945 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.780711 4945 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.780721 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.780729 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzmhx\" (UniqueName: \"kubernetes.io/projected/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-kube-api-access-nzmhx\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.780740 4945 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.780748 4945 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:05 crc kubenswrapper[4945]: I1008 15:28:05.780756 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6df6697f-bec8-4b87-a0ef-7e71c9a0b360-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.181585 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kr5ql_6df6697f-bec8-4b87-a0ef-7e71c9a0b360/console/0.log" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.181647 4945 generic.go:334] "Generic (PLEG): container finished" podID="6df6697f-bec8-4b87-a0ef-7e71c9a0b360" containerID="ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e" exitCode=2 Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.181676 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kr5ql" event={"ID":"6df6697f-bec8-4b87-a0ef-7e71c9a0b360","Type":"ContainerDied","Data":"ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e"} Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.181717 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kr5ql" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.181733 4945 scope.go:117] "RemoveContainer" containerID="ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.181723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kr5ql" event={"ID":"6df6697f-bec8-4b87-a0ef-7e71c9a0b360","Type":"ContainerDied","Data":"ca92baacaa8eaf12592e80a2ff5dcea43a3e725bc853f2be507c362ea791904c"} Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.204021 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kr5ql"] Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.206483 4945 scope.go:117] "RemoveContainer" containerID="ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e" Oct 08 15:28:06 crc kubenswrapper[4945]: E1008 15:28:06.207071 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e\": container with ID starting with ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e not found: ID does not exist" containerID="ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.207156 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e"} err="failed to get container status \"ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e\": rpc error: code = NotFound desc = could not find container \"ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e\": container with ID starting with ded2888691a30d77007c72df12f91254c6bfd002e9461c08989244a3278df64e not found: ID does not exist" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.209867 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-kr5ql"] Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.803847 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl"] Oct 08 15:28:06 crc kubenswrapper[4945]: E1008 15:28:06.804386 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df6697f-bec8-4b87-a0ef-7e71c9a0b360" containerName="console" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.804406 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df6697f-bec8-4b87-a0ef-7e71c9a0b360" containerName="console" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.804513 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df6697f-bec8-4b87-a0ef-7e71c9a0b360" containerName="console" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.805365 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.808900 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.814784 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl"] Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.895790 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpj59\" (UniqueName: \"kubernetes.io/projected/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-kube-api-access-gpj59\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.895876 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.895908 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.997302 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.997364 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.997448 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpj59\" (UniqueName: \"kubernetes.io/projected/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-kube-api-access-gpj59\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.997942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:06 crc kubenswrapper[4945]: I1008 15:28:06.997989 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:07 crc kubenswrapper[4945]: I1008 15:28:07.012598 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpj59\" (UniqueName: \"kubernetes.io/projected/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-kube-api-access-gpj59\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:07 crc kubenswrapper[4945]: I1008 15:28:07.129260 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:07 crc kubenswrapper[4945]: I1008 15:28:07.192328 4945 generic.go:334] "Generic (PLEG): container finished" podID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerID="d137a14ac5a92fffa1ca96e0676994ec4c2b84f4766e5c38ecd673c3071b1667" exitCode=0 Oct 08 15:28:07 crc kubenswrapper[4945]: I1008 15:28:07.192363 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9tgz" event={"ID":"207715ea-7f62-46c7-8351-f56bd1750ecd","Type":"ContainerDied","Data":"d137a14ac5a92fffa1ca96e0676994ec4c2b84f4766e5c38ecd673c3071b1667"} Oct 08 15:28:07 crc kubenswrapper[4945]: I1008 15:28:07.523502 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl"] Oct 08 15:28:08 crc kubenswrapper[4945]: I1008 15:28:08.031321 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df6697f-bec8-4b87-a0ef-7e71c9a0b360" path="/var/lib/kubelet/pods/6df6697f-bec8-4b87-a0ef-7e71c9a0b360/volumes" Oct 08 15:28:08 crc kubenswrapper[4945]: I1008 15:28:08.200684 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9tgz" event={"ID":"207715ea-7f62-46c7-8351-f56bd1750ecd","Type":"ContainerStarted","Data":"d9b55c372ccba84944a52600cd95bba888f3cd2db4c86c7d220c0a759c29ed6e"} Oct 08 15:28:08 crc kubenswrapper[4945]: I1008 15:28:08.202603 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerID="cd867d3de865c047f5748ba82b9a917a65d558a8288728b5aaf993d3c838d805" exitCode=0 Oct 08 15:28:08 crc kubenswrapper[4945]: I1008 15:28:08.202655 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" event={"ID":"b6c23df0-35e1-4905-bf5a-84b240ea3b0d","Type":"ContainerDied","Data":"cd867d3de865c047f5748ba82b9a917a65d558a8288728b5aaf993d3c838d805"} Oct 08 15:28:08 crc kubenswrapper[4945]: I1008 15:28:08.202694 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" event={"ID":"b6c23df0-35e1-4905-bf5a-84b240ea3b0d","Type":"ContainerStarted","Data":"c4ba0271a93647d529e34ac0224708c545010a099c5e2fea3127b51f133c5c2d"} Oct 08 15:28:08 crc kubenswrapper[4945]: I1008 15:28:08.219289 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p9tgz" podStartSLOduration=2.757692806 podStartE2EDuration="5.219270946s" podCreationTimestamp="2025-10-08 15:28:03 +0000 UTC" firstStartedPulling="2025-10-08 15:28:05.175257639 +0000 UTC m=+894.529172540" lastFinishedPulling="2025-10-08 15:28:07.636835779 +0000 UTC m=+896.990750680" observedRunningTime="2025-10-08 15:28:08.218211327 +0000 UTC m=+897.572126228" watchObservedRunningTime="2025-10-08 15:28:08.219270946 +0000 UTC m=+897.573185847" Oct 08 15:28:12 crc kubenswrapper[4945]: I1008 15:28:12.229774 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerID="13f8c7ce864066167d37cfa587fb6b4112e54e78af5c715ca16b74cd8a5cc818" exitCode=0 Oct 08 15:28:12 crc kubenswrapper[4945]: I1008 15:28:12.229942 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" event={"ID":"b6c23df0-35e1-4905-bf5a-84b240ea3b0d","Type":"ContainerDied","Data":"13f8c7ce864066167d37cfa587fb6b4112e54e78af5c715ca16b74cd8a5cc818"} Oct 08 15:28:13 crc kubenswrapper[4945]: I1008 15:28:13.238751 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" event={"ID":"b6c23df0-35e1-4905-bf5a-84b240ea3b0d","Type":"ContainerStarted","Data":"d27051b14097c92fc2cd25e8021d1db2516f164deab6ee61a18c1736a3a8303a"} Oct 08 15:28:13 crc kubenswrapper[4945]: I1008 15:28:13.670261 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:13 crc kubenswrapper[4945]: I1008 15:28:13.670325 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:13 crc kubenswrapper[4945]: I1008 15:28:13.712490 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:14 crc kubenswrapper[4945]: I1008 15:28:14.246362 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerID="d27051b14097c92fc2cd25e8021d1db2516f164deab6ee61a18c1736a3a8303a" exitCode=0 Oct 08 15:28:14 crc kubenswrapper[4945]: I1008 15:28:14.246581 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" event={"ID":"b6c23df0-35e1-4905-bf5a-84b240ea3b0d","Type":"ContainerDied","Data":"d27051b14097c92fc2cd25e8021d1db2516f164deab6ee61a18c1736a3a8303a"} Oct 08 15:28:14 crc kubenswrapper[4945]: I1008 15:28:14.291567 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.567706 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.612057 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-util\") pod \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.612245 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpj59\" (UniqueName: \"kubernetes.io/projected/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-kube-api-access-gpj59\") pod \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.612311 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-bundle\") pod \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\" (UID: \"b6c23df0-35e1-4905-bf5a-84b240ea3b0d\") " Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.613855 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-bundle" (OuterVolumeSpecName: "bundle") pod "b6c23df0-35e1-4905-bf5a-84b240ea3b0d" (UID: "b6c23df0-35e1-4905-bf5a-84b240ea3b0d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.621551 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-kube-api-access-gpj59" (OuterVolumeSpecName: "kube-api-access-gpj59") pod "b6c23df0-35e1-4905-bf5a-84b240ea3b0d" (UID: "b6c23df0-35e1-4905-bf5a-84b240ea3b0d"). InnerVolumeSpecName "kube-api-access-gpj59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.622896 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-util" (OuterVolumeSpecName: "util") pod "b6c23df0-35e1-4905-bf5a-84b240ea3b0d" (UID: "b6c23df0-35e1-4905-bf5a-84b240ea3b0d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.714299 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-util\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.714386 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpj59\" (UniqueName: \"kubernetes.io/projected/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-kube-api-access-gpj59\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:15 crc kubenswrapper[4945]: I1008 15:28:15.714413 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c23df0-35e1-4905-bf5a-84b240ea3b0d-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:16 crc kubenswrapper[4945]: I1008 15:28:16.263804 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" event={"ID":"b6c23df0-35e1-4905-bf5a-84b240ea3b0d","Type":"ContainerDied","Data":"c4ba0271a93647d529e34ac0224708c545010a099c5e2fea3127b51f133c5c2d"} Oct 08 15:28:16 crc kubenswrapper[4945]: I1008 15:28:16.263852 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4ba0271a93647d529e34ac0224708c545010a099c5e2fea3127b51f133c5c2d" Oct 08 15:28:16 crc kubenswrapper[4945]: I1008 15:28:16.263880 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl" Oct 08 15:28:17 crc kubenswrapper[4945]: I1008 15:28:17.925587 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9tgz"] Oct 08 15:28:17 crc kubenswrapper[4945]: I1008 15:28:17.926151 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p9tgz" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="registry-server" containerID="cri-o://d9b55c372ccba84944a52600cd95bba888f3cd2db4c86c7d220c0a759c29ed6e" gracePeriod=2 Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.275853 4945 generic.go:334] "Generic (PLEG): container finished" podID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerID="d9b55c372ccba84944a52600cd95bba888f3cd2db4c86c7d220c0a759c29ed6e" exitCode=0 Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.275895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9tgz" event={"ID":"207715ea-7f62-46c7-8351-f56bd1750ecd","Type":"ContainerDied","Data":"d9b55c372ccba84944a52600cd95bba888f3cd2db4c86c7d220c0a759c29ed6e"} Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.338170 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.451694 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw8zg\" (UniqueName: \"kubernetes.io/projected/207715ea-7f62-46c7-8351-f56bd1750ecd-kube-api-access-pw8zg\") pod \"207715ea-7f62-46c7-8351-f56bd1750ecd\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.452043 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-utilities\") pod \"207715ea-7f62-46c7-8351-f56bd1750ecd\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.452676 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-utilities" (OuterVolumeSpecName: "utilities") pod "207715ea-7f62-46c7-8351-f56bd1750ecd" (UID: "207715ea-7f62-46c7-8351-f56bd1750ecd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.452738 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-catalog-content\") pod \"207715ea-7f62-46c7-8351-f56bd1750ecd\" (UID: \"207715ea-7f62-46c7-8351-f56bd1750ecd\") " Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.452985 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.457288 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207715ea-7f62-46c7-8351-f56bd1750ecd-kube-api-access-pw8zg" (OuterVolumeSpecName: "kube-api-access-pw8zg") pod "207715ea-7f62-46c7-8351-f56bd1750ecd" (UID: "207715ea-7f62-46c7-8351-f56bd1750ecd"). InnerVolumeSpecName "kube-api-access-pw8zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.492607 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "207715ea-7f62-46c7-8351-f56bd1750ecd" (UID: "207715ea-7f62-46c7-8351-f56bd1750ecd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.553889 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/207715ea-7f62-46c7-8351-f56bd1750ecd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:18 crc kubenswrapper[4945]: I1008 15:28:18.553934 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw8zg\" (UniqueName: \"kubernetes.io/projected/207715ea-7f62-46c7-8351-f56bd1750ecd-kube-api-access-pw8zg\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:19 crc kubenswrapper[4945]: I1008 15:28:19.283205 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9tgz" event={"ID":"207715ea-7f62-46c7-8351-f56bd1750ecd","Type":"ContainerDied","Data":"d28d5de9c4acd4e05d70e045fdaf5f10b98a8152635df13f051d610d10585b0c"} Oct 08 15:28:19 crc kubenswrapper[4945]: I1008 15:28:19.283268 4945 scope.go:117] "RemoveContainer" containerID="d9b55c372ccba84944a52600cd95bba888f3cd2db4c86c7d220c0a759c29ed6e" Oct 08 15:28:19 crc kubenswrapper[4945]: I1008 15:28:19.283321 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9tgz" Oct 08 15:28:19 crc kubenswrapper[4945]: I1008 15:28:19.298834 4945 scope.go:117] "RemoveContainer" containerID="d137a14ac5a92fffa1ca96e0676994ec4c2b84f4766e5c38ecd673c3071b1667" Oct 08 15:28:19 crc kubenswrapper[4945]: I1008 15:28:19.319680 4945 scope.go:117] "RemoveContainer" containerID="88b8b91c9a7c33a2e7391012d95fb7243be148c6f92440a2936557d0da2d31bd" Oct 08 15:28:19 crc kubenswrapper[4945]: I1008 15:28:19.324283 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9tgz"] Oct 08 15:28:19 crc kubenswrapper[4945]: I1008 15:28:19.329276 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p9tgz"] Oct 08 15:28:20 crc kubenswrapper[4945]: I1008 15:28:20.031164 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" path="/var/lib/kubelet/pods/207715ea-7f62-46c7-8351-f56bd1750ecd/volumes" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.555585 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-575bbc8666-4t674"] Oct 08 15:28:23 crc kubenswrapper[4945]: E1008 15:28:23.556410 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="registry-server" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556421 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="registry-server" Oct 08 15:28:23 crc kubenswrapper[4945]: E1008 15:28:23.556432 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="extract-content" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556438 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="extract-content" Oct 08 15:28:23 crc kubenswrapper[4945]: E1008 15:28:23.556444 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerName="extract" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556450 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerName="extract" Oct 08 15:28:23 crc kubenswrapper[4945]: E1008 15:28:23.556459 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerName="util" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556464 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerName="util" Oct 08 15:28:23 crc kubenswrapper[4945]: E1008 15:28:23.556474 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="extract-utilities" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556480 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="extract-utilities" Oct 08 15:28:23 crc kubenswrapper[4945]: E1008 15:28:23.556510 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerName="pull" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556516 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerName="pull" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556606 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="207715ea-7f62-46c7-8351-f56bd1750ecd" containerName="registry-server" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556616 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c23df0-35e1-4905-bf5a-84b240ea3b0d" containerName="extract" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.556974 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.559673 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.559832 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.560369 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.560401 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-d6qqd" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.563091 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.592591 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-575bbc8666-4t674"] Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.613322 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13c15125-a872-459e-ac9a-d37356f47418-apiservice-cert\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.613430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzgfr\" (UniqueName: \"kubernetes.io/projected/13c15125-a872-459e-ac9a-d37356f47418-kube-api-access-rzgfr\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.613475 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13c15125-a872-459e-ac9a-d37356f47418-webhook-cert\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.714807 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13c15125-a872-459e-ac9a-d37356f47418-apiservice-cert\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.714890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzgfr\" (UniqueName: \"kubernetes.io/projected/13c15125-a872-459e-ac9a-d37356f47418-kube-api-access-rzgfr\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.714912 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13c15125-a872-459e-ac9a-d37356f47418-webhook-cert\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.720680 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13c15125-a872-459e-ac9a-d37356f47418-webhook-cert\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.720721 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13c15125-a872-459e-ac9a-d37356f47418-apiservice-cert\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.735402 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzgfr\" (UniqueName: \"kubernetes.io/projected/13c15125-a872-459e-ac9a-d37356f47418-kube-api-access-rzgfr\") pod \"metallb-operator-controller-manager-575bbc8666-4t674\" (UID: \"13c15125-a872-459e-ac9a-d37356f47418\") " pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.790409 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x"] Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.791103 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.794991 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.795364 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.795447 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-7vm92" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.850394 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x"] Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.871004 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.917481 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/132bc709-76d1-41fc-ae20-04141e80a6ac-webhook-cert\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.917561 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tclct\" (UniqueName: \"kubernetes.io/projected/132bc709-76d1-41fc-ae20-04141e80a6ac-kube-api-access-tclct\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:23 crc kubenswrapper[4945]: I1008 15:28:23.917589 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/132bc709-76d1-41fc-ae20-04141e80a6ac-apiservice-cert\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.018478 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/132bc709-76d1-41fc-ae20-04141e80a6ac-webhook-cert\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.018823 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tclct\" (UniqueName: \"kubernetes.io/projected/132bc709-76d1-41fc-ae20-04141e80a6ac-kube-api-access-tclct\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.018853 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/132bc709-76d1-41fc-ae20-04141e80a6ac-apiservice-cert\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.025373 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/132bc709-76d1-41fc-ae20-04141e80a6ac-apiservice-cert\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.025846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/132bc709-76d1-41fc-ae20-04141e80a6ac-webhook-cert\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.039138 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tclct\" (UniqueName: \"kubernetes.io/projected/132bc709-76d1-41fc-ae20-04141e80a6ac-kube-api-access-tclct\") pod \"metallb-operator-webhook-server-7d999db554-m5t6x\" (UID: \"132bc709-76d1-41fc-ae20-04141e80a6ac\") " pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.108649 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.336064 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x"] Oct 08 15:28:24 crc kubenswrapper[4945]: W1008 15:28:24.346326 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod132bc709_76d1_41fc_ae20_04141e80a6ac.slice/crio-353db1158135c14c3e0ca4f74b8e742dad798e3b35bd5362663efc924fd0a69b WatchSource:0}: Error finding container 353db1158135c14c3e0ca4f74b8e742dad798e3b35bd5362663efc924fd0a69b: Status 404 returned error can't find the container with id 353db1158135c14c3e0ca4f74b8e742dad798e3b35bd5362663efc924fd0a69b Oct 08 15:28:24 crc kubenswrapper[4945]: I1008 15:28:24.373180 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-575bbc8666-4t674"] Oct 08 15:28:24 crc kubenswrapper[4945]: W1008 15:28:24.384411 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13c15125_a872_459e_ac9a_d37356f47418.slice/crio-cc55563dc8726db6ad74586ba365f36b868b522830721fda23b5659f67adf637 WatchSource:0}: Error finding container cc55563dc8726db6ad74586ba365f36b868b522830721fda23b5659f67adf637: Status 404 returned error can't find the container with id cc55563dc8726db6ad74586ba365f36b868b522830721fda23b5659f67adf637 Oct 08 15:28:25 crc kubenswrapper[4945]: I1008 15:28:25.321410 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" event={"ID":"13c15125-a872-459e-ac9a-d37356f47418","Type":"ContainerStarted","Data":"cc55563dc8726db6ad74586ba365f36b868b522830721fda23b5659f67adf637"} Oct 08 15:28:25 crc kubenswrapper[4945]: I1008 15:28:25.322826 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" event={"ID":"132bc709-76d1-41fc-ae20-04141e80a6ac","Type":"ContainerStarted","Data":"353db1158135c14c3e0ca4f74b8e742dad798e3b35bd5362663efc924fd0a69b"} Oct 08 15:28:29 crc kubenswrapper[4945]: I1008 15:28:29.348140 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" event={"ID":"13c15125-a872-459e-ac9a-d37356f47418","Type":"ContainerStarted","Data":"91fc94ab4232ae02f091db89a800840e0356bdda564b1c1290173a1067374757"} Oct 08 15:28:29 crc kubenswrapper[4945]: I1008 15:28:29.348543 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:28:29 crc kubenswrapper[4945]: I1008 15:28:29.370648 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" podStartSLOduration=3.169028118 podStartE2EDuration="6.370624182s" podCreationTimestamp="2025-10-08 15:28:23 +0000 UTC" firstStartedPulling="2025-10-08 15:28:24.392256115 +0000 UTC m=+913.746171016" lastFinishedPulling="2025-10-08 15:28:27.593852179 +0000 UTC m=+916.947767080" observedRunningTime="2025-10-08 15:28:29.365881939 +0000 UTC m=+918.719796860" watchObservedRunningTime="2025-10-08 15:28:29.370624182 +0000 UTC m=+918.724539093" Oct 08 15:28:30 crc kubenswrapper[4945]: I1008 15:28:30.354272 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" event={"ID":"132bc709-76d1-41fc-ae20-04141e80a6ac","Type":"ContainerStarted","Data":"ea2b728935eb7b143e2129f8e48e8f52aa5561bbdd4aca0edce3700e358668ed"} Oct 08 15:28:31 crc kubenswrapper[4945]: I1008 15:28:31.358988 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:44 crc kubenswrapper[4945]: I1008 15:28:44.114133 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" Oct 08 15:28:44 crc kubenswrapper[4945]: I1008 15:28:44.132629 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7d999db554-m5t6x" podStartSLOduration=16.100874732 podStartE2EDuration="21.132610662s" podCreationTimestamp="2025-10-08 15:28:23 +0000 UTC" firstStartedPulling="2025-10-08 15:28:24.362303579 +0000 UTC m=+913.716218480" lastFinishedPulling="2025-10-08 15:28:29.394039499 +0000 UTC m=+918.747954410" observedRunningTime="2025-10-08 15:28:30.374476648 +0000 UTC m=+919.728391549" watchObservedRunningTime="2025-10-08 15:28:44.132610662 +0000 UTC m=+933.486525563" Oct 08 15:28:49 crc kubenswrapper[4945]: I1008 15:28:49.184232 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:28:49 crc kubenswrapper[4945]: I1008 15:28:49.184741 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:29:03 crc kubenswrapper[4945]: I1008 15:29:03.874657 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-575bbc8666-4t674" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.670649 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-cc5ps"] Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.673261 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.676437 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.676668 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.679812 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-5vmzd" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.680825 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z"] Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.681501 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.685504 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.695191 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z"] Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.767473 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-dftb8"] Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768338 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-h5s5z\" (UID: \"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768469 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-sockets\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768550 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh9lf\" (UniqueName: \"kubernetes.io/projected/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-kube-api-access-sh9lf\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768629 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-metrics-certs\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-startup\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768774 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-conf\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768845 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-reloader\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.768927 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-metrics\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.769012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hqzc\" (UniqueName: \"kubernetes.io/projected/1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c-kube-api-access-5hqzc\") pod \"frr-k8s-webhook-server-64bf5d555-h5s5z\" (UID: \"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.769122 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.773005 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.773061 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.773278 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-sw2ss" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.773361 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.787980 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-nnbmf"] Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.790009 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.793520 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.802128 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-nnbmf"] Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870460 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/919b5756-8417-47d1-acd7-fbfd2c01d246-cert\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870540 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-h5s5z\" (UID: \"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870577 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870635 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/283c80b9-bfe4-4c26-852d-9ba6d14009d4-metallb-excludel2\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870664 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-sockets\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870709 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjdrm\" (UniqueName: \"kubernetes.io/projected/919b5756-8417-47d1-acd7-fbfd2c01d246-kube-api-access-cjdrm\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870735 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh9lf\" (UniqueName: \"kubernetes.io/projected/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-kube-api-access-sh9lf\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-startup\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870796 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-metrics-certs\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870821 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtxhg\" (UniqueName: \"kubernetes.io/projected/283c80b9-bfe4-4c26-852d-9ba6d14009d4-kube-api-access-qtxhg\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870868 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-reloader\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870889 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-conf\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870930 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-metrics-certs\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.870963 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-metrics\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.871007 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/919b5756-8417-47d1-acd7-fbfd2c01d246-metrics-certs\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.871046 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hqzc\" (UniqueName: \"kubernetes.io/projected/1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c-kube-api-access-5hqzc\") pod \"frr-k8s-webhook-server-64bf5d555-h5s5z\" (UID: \"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.871648 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-conf\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.872067 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-startup\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.872447 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-metrics\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.872482 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-frr-sockets\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.872600 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-reloader\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.878857 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-metrics-certs\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.879826 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-h5s5z\" (UID: \"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.891853 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hqzc\" (UniqueName: \"kubernetes.io/projected/1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c-kube-api-access-5hqzc\") pod \"frr-k8s-webhook-server-64bf5d555-h5s5z\" (UID: \"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.897048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh9lf\" (UniqueName: \"kubernetes.io/projected/c0c94cdf-157a-48b2-b8ff-389b2ec3e04d-kube-api-access-sh9lf\") pod \"frr-k8s-cc5ps\" (UID: \"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d\") " pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.972638 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/283c80b9-bfe4-4c26-852d-9ba6d14009d4-metallb-excludel2\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.972694 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjdrm\" (UniqueName: \"kubernetes.io/projected/919b5756-8417-47d1-acd7-fbfd2c01d246-kube-api-access-cjdrm\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.972725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtxhg\" (UniqueName: \"kubernetes.io/projected/283c80b9-bfe4-4c26-852d-9ba6d14009d4-kube-api-access-qtxhg\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.972758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-metrics-certs\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.972795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/919b5756-8417-47d1-acd7-fbfd2c01d246-metrics-certs\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.972863 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/919b5756-8417-47d1-acd7-fbfd2c01d246-cert\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.972895 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: E1008 15:29:04.973031 4945 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 15:29:04 crc kubenswrapper[4945]: E1008 15:29:04.973098 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist podName:283c80b9-bfe4-4c26-852d-9ba6d14009d4 nodeName:}" failed. No retries permitted until 2025-10-08 15:29:05.473078967 +0000 UTC m=+954.826993868 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist") pod "speaker-dftb8" (UID: "283c80b9-bfe4-4c26-852d-9ba6d14009d4") : secret "metallb-memberlist" not found Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.973472 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/283c80b9-bfe4-4c26-852d-9ba6d14009d4-metallb-excludel2\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.977660 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/919b5756-8417-47d1-acd7-fbfd2c01d246-metrics-certs\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.977776 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/919b5756-8417-47d1-acd7-fbfd2c01d246-cert\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.979040 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-metrics-certs\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:04 crc kubenswrapper[4945]: I1008 15:29:04.998589 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtxhg\" (UniqueName: \"kubernetes.io/projected/283c80b9-bfe4-4c26-852d-9ba6d14009d4-kube-api-access-qtxhg\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.002182 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.011105 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjdrm\" (UniqueName: \"kubernetes.io/projected/919b5756-8417-47d1-acd7-fbfd2c01d246-kube-api-access-cjdrm\") pod \"controller-68d546b9d8-nnbmf\" (UID: \"919b5756-8417-47d1-acd7-fbfd2c01d246\") " pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.011425 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.119787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.428728 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z"] Oct 08 15:29:05 crc kubenswrapper[4945]: W1008 15:29:05.431699 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ae8dfbc_86f8_4b14_a8a3_42d6e7f12b3c.slice/crio-0139a3292d55bc443833d3e2d9516f96be71fd1bc5ec9c3fcc2137f7445d7ee1 WatchSource:0}: Error finding container 0139a3292d55bc443833d3e2d9516f96be71fd1bc5ec9c3fcc2137f7445d7ee1: Status 404 returned error can't find the container with id 0139a3292d55bc443833d3e2d9516f96be71fd1bc5ec9c3fcc2137f7445d7ee1 Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.479659 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:05 crc kubenswrapper[4945]: E1008 15:29:05.479798 4945 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 15:29:05 crc kubenswrapper[4945]: E1008 15:29:05.479853 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist podName:283c80b9-bfe4-4c26-852d-9ba6d14009d4 nodeName:}" failed. No retries permitted until 2025-10-08 15:29:06.479837045 +0000 UTC m=+955.833751946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist") pod "speaker-dftb8" (UID: "283c80b9-bfe4-4c26-852d-9ba6d14009d4") : secret "metallb-memberlist" not found Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.551492 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-nnbmf"] Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.565505 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" event={"ID":"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c","Type":"ContainerStarted","Data":"0139a3292d55bc443833d3e2d9516f96be71fd1bc5ec9c3fcc2137f7445d7ee1"} Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.567325 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-nnbmf" event={"ID":"919b5756-8417-47d1-acd7-fbfd2c01d246","Type":"ContainerStarted","Data":"251de5e33246864fd481d2ab8d7c78eba3abf117563018ad8a903848575299c7"} Oct 08 15:29:05 crc kubenswrapper[4945]: I1008 15:29:05.568475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerStarted","Data":"262da17dde8444145d06a115314aa012e9d7641c0ba2b4caf758239d0907a59b"} Oct 08 15:29:06 crc kubenswrapper[4945]: I1008 15:29:06.493933 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:06 crc kubenswrapper[4945]: I1008 15:29:06.499969 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283c80b9-bfe4-4c26-852d-9ba6d14009d4-memberlist\") pod \"speaker-dftb8\" (UID: \"283c80b9-bfe4-4c26-852d-9ba6d14009d4\") " pod="metallb-system/speaker-dftb8" Oct 08 15:29:06 crc kubenswrapper[4945]: I1008 15:29:06.577402 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-nnbmf" event={"ID":"919b5756-8417-47d1-acd7-fbfd2c01d246","Type":"ContainerStarted","Data":"5a1a66cde3e2b88fb2003c79cf23748c18a829a531e9308cc3d19220008da2e0"} Oct 08 15:29:06 crc kubenswrapper[4945]: I1008 15:29:06.577443 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-nnbmf" event={"ID":"919b5756-8417-47d1-acd7-fbfd2c01d246","Type":"ContainerStarted","Data":"c74756173669948f910abad5adaa34a783726f672a0ec4a5a373e085d6914fd7"} Oct 08 15:29:06 crc kubenswrapper[4945]: I1008 15:29:06.577552 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:06 crc kubenswrapper[4945]: I1008 15:29:06.597460 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dftb8" Oct 08 15:29:06 crc kubenswrapper[4945]: W1008 15:29:06.630632 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod283c80b9_bfe4_4c26_852d_9ba6d14009d4.slice/crio-b0b146c6224518554fc4c4cfeb57b1bd75a2845bdc4f2e421d978f4ca6fcfc65 WatchSource:0}: Error finding container b0b146c6224518554fc4c4cfeb57b1bd75a2845bdc4f2e421d978f4ca6fcfc65: Status 404 returned error can't find the container with id b0b146c6224518554fc4c4cfeb57b1bd75a2845bdc4f2e421d978f4ca6fcfc65 Oct 08 15:29:07 crc kubenswrapper[4945]: I1008 15:29:07.590686 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dftb8" event={"ID":"283c80b9-bfe4-4c26-852d-9ba6d14009d4","Type":"ContainerStarted","Data":"1386d9829f1bc9c550682ba47f4c2cc06fb74cf7c35a1bc5fe601b797e4cb7d4"} Oct 08 15:29:07 crc kubenswrapper[4945]: I1008 15:29:07.591022 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dftb8" event={"ID":"283c80b9-bfe4-4c26-852d-9ba6d14009d4","Type":"ContainerStarted","Data":"f488801e7ba209000a070ab107a0dbb7e85ce72c065efda4d743d698c01b9307"} Oct 08 15:29:07 crc kubenswrapper[4945]: I1008 15:29:07.591032 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dftb8" event={"ID":"283c80b9-bfe4-4c26-852d-9ba6d14009d4","Type":"ContainerStarted","Data":"b0b146c6224518554fc4c4cfeb57b1bd75a2845bdc4f2e421d978f4ca6fcfc65"} Oct 08 15:29:07 crc kubenswrapper[4945]: I1008 15:29:07.591260 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-dftb8" Oct 08 15:29:07 crc kubenswrapper[4945]: I1008 15:29:07.605710 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-nnbmf" podStartSLOduration=3.6056922 podStartE2EDuration="3.6056922s" podCreationTimestamp="2025-10-08 15:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:29:06.629263965 +0000 UTC m=+955.983178866" watchObservedRunningTime="2025-10-08 15:29:07.6056922 +0000 UTC m=+956.959607101" Oct 08 15:29:07 crc kubenswrapper[4945]: I1008 15:29:07.609469 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-dftb8" podStartSLOduration=3.609454116 podStartE2EDuration="3.609454116s" podCreationTimestamp="2025-10-08 15:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:29:07.6053571 +0000 UTC m=+956.959272001" watchObservedRunningTime="2025-10-08 15:29:07.609454116 +0000 UTC m=+956.963369027" Oct 08 15:29:13 crc kubenswrapper[4945]: I1008 15:29:13.649146 4945 generic.go:334] "Generic (PLEG): container finished" podID="c0c94cdf-157a-48b2-b8ff-389b2ec3e04d" containerID="eda4303462ee3f5bf27d9f3f2aab7851bc359999b8d5459bcb0be4ca823567c5" exitCode=0 Oct 08 15:29:13 crc kubenswrapper[4945]: I1008 15:29:13.649217 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerDied","Data":"eda4303462ee3f5bf27d9f3f2aab7851bc359999b8d5459bcb0be4ca823567c5"} Oct 08 15:29:13 crc kubenswrapper[4945]: I1008 15:29:13.653666 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" event={"ID":"1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c","Type":"ContainerStarted","Data":"6bc874c3b96abd68a674f0b5219ba495863f599c69c83ee267d89b6512c38de1"} Oct 08 15:29:13 crc kubenswrapper[4945]: I1008 15:29:13.653937 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:14 crc kubenswrapper[4945]: I1008 15:29:14.660883 4945 generic.go:334] "Generic (PLEG): container finished" podID="c0c94cdf-157a-48b2-b8ff-389b2ec3e04d" containerID="7f615a9eb6e457b1df576997f1362b778f8f08cc4d66ebb35b1f3b8223f5a66b" exitCode=0 Oct 08 15:29:14 crc kubenswrapper[4945]: I1008 15:29:14.660929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerDied","Data":"7f615a9eb6e457b1df576997f1362b778f8f08cc4d66ebb35b1f3b8223f5a66b"} Oct 08 15:29:14 crc kubenswrapper[4945]: I1008 15:29:14.684334 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" podStartSLOduration=2.863053326 podStartE2EDuration="10.684316005s" podCreationTimestamp="2025-10-08 15:29:04 +0000 UTC" firstStartedPulling="2025-10-08 15:29:05.434911621 +0000 UTC m=+954.788826522" lastFinishedPulling="2025-10-08 15:29:13.2561743 +0000 UTC m=+962.610089201" observedRunningTime="2025-10-08 15:29:13.692986144 +0000 UTC m=+963.046901095" watchObservedRunningTime="2025-10-08 15:29:14.684316005 +0000 UTC m=+964.038230906" Oct 08 15:29:15 crc kubenswrapper[4945]: I1008 15:29:15.123784 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-nnbmf" Oct 08 15:29:15 crc kubenswrapper[4945]: I1008 15:29:15.670006 4945 generic.go:334] "Generic (PLEG): container finished" podID="c0c94cdf-157a-48b2-b8ff-389b2ec3e04d" containerID="ff2c4eb5fb1d41025af9a8a8386922216b772970154b3e74edcfb934bb0fc4c6" exitCode=0 Oct 08 15:29:15 crc kubenswrapper[4945]: I1008 15:29:15.670070 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerDied","Data":"ff2c4eb5fb1d41025af9a8a8386922216b772970154b3e74edcfb934bb0fc4c6"} Oct 08 15:29:16 crc kubenswrapper[4945]: I1008 15:29:16.603867 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-dftb8" Oct 08 15:29:16 crc kubenswrapper[4945]: I1008 15:29:16.690121 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerStarted","Data":"d18b539c1fc3935f78a422e87206ae9d20966054cd07967f991657c6d889f063"} Oct 08 15:29:16 crc kubenswrapper[4945]: I1008 15:29:16.690155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerStarted","Data":"22c7be4fc24e1d2da096d9f3cd306d2dea724b4ecc3296cd138cc3104de64f37"} Oct 08 15:29:16 crc kubenswrapper[4945]: I1008 15:29:16.690165 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerStarted","Data":"3611d07a6201c3fe33de782356d0d59fd3573d751a6b4b2318bfddcb4e5ca4b1"} Oct 08 15:29:16 crc kubenswrapper[4945]: I1008 15:29:16.690174 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerStarted","Data":"50c16a8bdab4c19aa6c015b67dded5f6dc33816dad774469f8feca34d8166233"} Oct 08 15:29:17 crc kubenswrapper[4945]: I1008 15:29:17.702607 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerStarted","Data":"44fe53eb7600420c307fd2c06f8af43933113d9306c60f5c9c5223a0cdcdb49b"} Oct 08 15:29:17 crc kubenswrapper[4945]: I1008 15:29:17.703088 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:17 crc kubenswrapper[4945]: I1008 15:29:17.703134 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cc5ps" event={"ID":"c0c94cdf-157a-48b2-b8ff-389b2ec3e04d","Type":"ContainerStarted","Data":"43abdfcb704a61ed9f52312db25d49c186e556fad4b8459224b13da46bebd4c5"} Oct 08 15:29:17 crc kubenswrapper[4945]: I1008 15:29:17.729875 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-cc5ps" podStartSLOduration=5.812970013 podStartE2EDuration="13.729853901s" podCreationTimestamp="2025-10-08 15:29:04 +0000 UTC" firstStartedPulling="2025-10-08 15:29:05.30950008 +0000 UTC m=+954.663414981" lastFinishedPulling="2025-10-08 15:29:13.226383968 +0000 UTC m=+962.580298869" observedRunningTime="2025-10-08 15:29:17.728990139 +0000 UTC m=+967.082905040" watchObservedRunningTime="2025-10-08 15:29:17.729853901 +0000 UTC m=+967.083768812" Oct 08 15:29:19 crc kubenswrapper[4945]: I1008 15:29:19.184801 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:29:19 crc kubenswrapper[4945]: I1008 15:29:19.186699 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:29:20 crc kubenswrapper[4945]: I1008 15:29:20.002682 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:20 crc kubenswrapper[4945]: I1008 15:29:20.048474 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.336161 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-65hg6"] Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.337753 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.340072 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.341285 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.345939 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-rqkkv" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.350303 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-65hg6"] Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.408314 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvq2h\" (UniqueName: \"kubernetes.io/projected/9844bfac-ebec-4692-abda-eacfe73d5b3a-kube-api-access-pvq2h\") pod \"openstack-operator-index-65hg6\" (UID: \"9844bfac-ebec-4692-abda-eacfe73d5b3a\") " pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.509148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvq2h\" (UniqueName: \"kubernetes.io/projected/9844bfac-ebec-4692-abda-eacfe73d5b3a-kube-api-access-pvq2h\") pod \"openstack-operator-index-65hg6\" (UID: \"9844bfac-ebec-4692-abda-eacfe73d5b3a\") " pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.530794 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvq2h\" (UniqueName: \"kubernetes.io/projected/9844bfac-ebec-4692-abda-eacfe73d5b3a-kube-api-access-pvq2h\") pod \"openstack-operator-index-65hg6\" (UID: \"9844bfac-ebec-4692-abda-eacfe73d5b3a\") " pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:23 crc kubenswrapper[4945]: I1008 15:29:23.663443 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:24 crc kubenswrapper[4945]: I1008 15:29:24.094048 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-65hg6"] Oct 08 15:29:24 crc kubenswrapper[4945]: I1008 15:29:24.753338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-65hg6" event={"ID":"9844bfac-ebec-4692-abda-eacfe73d5b3a","Type":"ContainerStarted","Data":"1486929d977fc13e8308764c4e1dc9fad7e4a707409b78f42f79b74e9b72b61a"} Oct 08 15:29:25 crc kubenswrapper[4945]: I1008 15:29:25.006097 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-cc5ps" Oct 08 15:29:25 crc kubenswrapper[4945]: I1008 15:29:25.034630 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-h5s5z" Oct 08 15:29:28 crc kubenswrapper[4945]: I1008 15:29:28.786181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-65hg6" event={"ID":"9844bfac-ebec-4692-abda-eacfe73d5b3a","Type":"ContainerStarted","Data":"ab5c9c1ec1a6ccacdac29bbcdaf5ef579f849590c74f93647f2f6afbcd59218e"} Oct 08 15:29:28 crc kubenswrapper[4945]: I1008 15:29:28.811233 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-65hg6" podStartSLOduration=2.042107305 podStartE2EDuration="5.81121298s" podCreationTimestamp="2025-10-08 15:29:23 +0000 UTC" firstStartedPulling="2025-10-08 15:29:24.1026902 +0000 UTC m=+973.456605091" lastFinishedPulling="2025-10-08 15:29:27.871795855 +0000 UTC m=+977.225710766" observedRunningTime="2025-10-08 15:29:28.808409887 +0000 UTC m=+978.162324788" watchObservedRunningTime="2025-10-08 15:29:28.81121298 +0000 UTC m=+978.165127881" Oct 08 15:29:33 crc kubenswrapper[4945]: I1008 15:29:33.664478 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:33 crc kubenswrapper[4945]: I1008 15:29:33.664865 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:33 crc kubenswrapper[4945]: I1008 15:29:33.703650 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:33 crc kubenswrapper[4945]: I1008 15:29:33.841876 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-65hg6" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.373625 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl"] Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.375985 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.378939 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6vq7j" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.387005 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl"] Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.488677 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm8cc\" (UniqueName: \"kubernetes.io/projected/00cf9b93-fcae-4759-936e-58397840df0a-kube-api-access-zm8cc\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.488752 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-util\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.488865 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-bundle\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.590602 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-bundle\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.590813 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm8cc\" (UniqueName: \"kubernetes.io/projected/00cf9b93-fcae-4759-936e-58397840df0a-kube-api-access-zm8cc\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.590862 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-util\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.591575 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-bundle\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.591704 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-util\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.617375 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm8cc\" (UniqueName: \"kubernetes.io/projected/00cf9b93-fcae-4759-936e-58397840df0a-kube-api-access-zm8cc\") pod \"a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:36 crc kubenswrapper[4945]: I1008 15:29:36.703606 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:37 crc kubenswrapper[4945]: I1008 15:29:37.139910 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl"] Oct 08 15:29:37 crc kubenswrapper[4945]: I1008 15:29:37.857163 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" event={"ID":"00cf9b93-fcae-4759-936e-58397840df0a","Type":"ContainerStarted","Data":"6d357d7b23e71903e672ad5a4845c91b152ae6523c4e3095321bee974b4c937d"} Oct 08 15:29:37 crc kubenswrapper[4945]: I1008 15:29:37.858227 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" event={"ID":"00cf9b93-fcae-4759-936e-58397840df0a","Type":"ContainerStarted","Data":"afcc5cc92bac1fa7d342d8712ab94c787a0fb10681c4e3e1b3aa97ff43da826c"} Oct 08 15:29:38 crc kubenswrapper[4945]: I1008 15:29:38.882186 4945 generic.go:334] "Generic (PLEG): container finished" podID="00cf9b93-fcae-4759-936e-58397840df0a" containerID="6d357d7b23e71903e672ad5a4845c91b152ae6523c4e3095321bee974b4c937d" exitCode=0 Oct 08 15:29:38 crc kubenswrapper[4945]: I1008 15:29:38.882257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" event={"ID":"00cf9b93-fcae-4759-936e-58397840df0a","Type":"ContainerDied","Data":"6d357d7b23e71903e672ad5a4845c91b152ae6523c4e3095321bee974b4c937d"} Oct 08 15:29:47 crc kubenswrapper[4945]: I1008 15:29:47.946244 4945 generic.go:334] "Generic (PLEG): container finished" podID="00cf9b93-fcae-4759-936e-58397840df0a" containerID="51e04f93b9e836ad2df73d7e75db026ff435a0baa3dc6edc9be2acaa93c4ecf1" exitCode=0 Oct 08 15:29:47 crc kubenswrapper[4945]: I1008 15:29:47.946335 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" event={"ID":"00cf9b93-fcae-4759-936e-58397840df0a","Type":"ContainerDied","Data":"51e04f93b9e836ad2df73d7e75db026ff435a0baa3dc6edc9be2acaa93c4ecf1"} Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.184781 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.185325 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.185403 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.186306 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06c6bd45fc6832af457ec9bd8f12ba1218de606e23dd6ce0bb0b843b24d0f361"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.186406 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://06c6bd45fc6832af457ec9bd8f12ba1218de606e23dd6ce0bb0b843b24d0f361" gracePeriod=600 Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.961194 4945 generic.go:334] "Generic (PLEG): container finished" podID="00cf9b93-fcae-4759-936e-58397840df0a" containerID="7c74abd54263c8dabaf9f34b57b371be9b5d78996f81a8e3f0ab51fa1ceea186" exitCode=0 Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.961364 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" event={"ID":"00cf9b93-fcae-4759-936e-58397840df0a","Type":"ContainerDied","Data":"7c74abd54263c8dabaf9f34b57b371be9b5d78996f81a8e3f0ab51fa1ceea186"} Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.968584 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="06c6bd45fc6832af457ec9bd8f12ba1218de606e23dd6ce0bb0b843b24d0f361" exitCode=0 Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.968645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"06c6bd45fc6832af457ec9bd8f12ba1218de606e23dd6ce0bb0b843b24d0f361"} Oct 08 15:29:49 crc kubenswrapper[4945]: I1008 15:29:49.968686 4945 scope.go:117] "RemoveContainer" containerID="ec80442af9d49173b57d5cd8aba7514d1075212c202e9f51e35b871dde889d11" Oct 08 15:29:50 crc kubenswrapper[4945]: I1008 15:29:50.977976 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"d35c98890277031b010217d13608d3edf99fe1fc8eaff9d1b2452633ef6288de"} Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.220654 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.299423 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-bundle\") pod \"00cf9b93-fcae-4759-936e-58397840df0a\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.299481 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm8cc\" (UniqueName: \"kubernetes.io/projected/00cf9b93-fcae-4759-936e-58397840df0a-kube-api-access-zm8cc\") pod \"00cf9b93-fcae-4759-936e-58397840df0a\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.299533 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-util\") pod \"00cf9b93-fcae-4759-936e-58397840df0a\" (UID: \"00cf9b93-fcae-4759-936e-58397840df0a\") " Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.300337 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-bundle" (OuterVolumeSpecName: "bundle") pod "00cf9b93-fcae-4759-936e-58397840df0a" (UID: "00cf9b93-fcae-4759-936e-58397840df0a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.307930 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00cf9b93-fcae-4759-936e-58397840df0a-kube-api-access-zm8cc" (OuterVolumeSpecName: "kube-api-access-zm8cc") pod "00cf9b93-fcae-4759-936e-58397840df0a" (UID: "00cf9b93-fcae-4759-936e-58397840df0a"). InnerVolumeSpecName "kube-api-access-zm8cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.314439 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-util" (OuterVolumeSpecName: "util") pod "00cf9b93-fcae-4759-936e-58397840df0a" (UID: "00cf9b93-fcae-4759-936e-58397840df0a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.401316 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.401351 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm8cc\" (UniqueName: \"kubernetes.io/projected/00cf9b93-fcae-4759-936e-58397840df0a-kube-api-access-zm8cc\") on node \"crc\" DevicePath \"\"" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.401365 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00cf9b93-fcae-4759-936e-58397840df0a-util\") on node \"crc\" DevicePath \"\"" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.988577 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.988564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl" event={"ID":"00cf9b93-fcae-4759-936e-58397840df0a","Type":"ContainerDied","Data":"afcc5cc92bac1fa7d342d8712ab94c787a0fb10681c4e3e1b3aa97ff43da826c"} Oct 08 15:29:51 crc kubenswrapper[4945]: I1008 15:29:51.988636 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afcc5cc92bac1fa7d342d8712ab94c787a0fb10681c4e3e1b3aa97ff43da826c" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.921282 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5"] Oct 08 15:29:58 crc kubenswrapper[4945]: E1008 15:29:58.922020 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cf9b93-fcae-4759-936e-58397840df0a" containerName="extract" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.922032 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cf9b93-fcae-4759-936e-58397840df0a" containerName="extract" Oct 08 15:29:58 crc kubenswrapper[4945]: E1008 15:29:58.922044 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cf9b93-fcae-4759-936e-58397840df0a" containerName="util" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.922050 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cf9b93-fcae-4759-936e-58397840df0a" containerName="util" Oct 08 15:29:58 crc kubenswrapper[4945]: E1008 15:29:58.922066 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cf9b93-fcae-4759-936e-58397840df0a" containerName="pull" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.922072 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cf9b93-fcae-4759-936e-58397840df0a" containerName="pull" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.922212 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="00cf9b93-fcae-4759-936e-58397840df0a" containerName="extract" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.922861 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.925954 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-npdsw" Oct 08 15:29:58 crc kubenswrapper[4945]: I1008 15:29:58.943186 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5"] Oct 08 15:29:59 crc kubenswrapper[4945]: I1008 15:29:59.014862 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwbqs\" (UniqueName: \"kubernetes.io/projected/ef574f9e-15f9-42fe-8593-5575c31fe0d9-kube-api-access-zwbqs\") pod \"openstack-operator-controller-operator-75967d884c-qslk5\" (UID: \"ef574f9e-15f9-42fe-8593-5575c31fe0d9\") " pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" Oct 08 15:29:59 crc kubenswrapper[4945]: I1008 15:29:59.116638 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwbqs\" (UniqueName: \"kubernetes.io/projected/ef574f9e-15f9-42fe-8593-5575c31fe0d9-kube-api-access-zwbqs\") pod \"openstack-operator-controller-operator-75967d884c-qslk5\" (UID: \"ef574f9e-15f9-42fe-8593-5575c31fe0d9\") " pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" Oct 08 15:29:59 crc kubenswrapper[4945]: I1008 15:29:59.134681 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwbqs\" (UniqueName: \"kubernetes.io/projected/ef574f9e-15f9-42fe-8593-5575c31fe0d9-kube-api-access-zwbqs\") pod \"openstack-operator-controller-operator-75967d884c-qslk5\" (UID: \"ef574f9e-15f9-42fe-8593-5575c31fe0d9\") " pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" Oct 08 15:29:59 crc kubenswrapper[4945]: I1008 15:29:59.239076 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" Oct 08 15:29:59 crc kubenswrapper[4945]: I1008 15:29:59.680583 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5"] Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.044195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" event={"ID":"ef574f9e-15f9-42fe-8593-5575c31fe0d9","Type":"ContainerStarted","Data":"83a64d465284cd7d88bf8e6569ab735cd84b7e032290930dddd69ea9f6687e19"} Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.139882 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t"] Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.140836 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.143936 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.148136 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.155545 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t"] Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.230965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a1c1792-e047-413d-8667-791d14563fb1-secret-volume\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.231030 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl2bx\" (UniqueName: \"kubernetes.io/projected/1a1c1792-e047-413d-8667-791d14563fb1-kube-api-access-tl2bx\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.231120 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a1c1792-e047-413d-8667-791d14563fb1-config-volume\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.334468 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a1c1792-e047-413d-8667-791d14563fb1-config-volume\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.334538 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a1c1792-e047-413d-8667-791d14563fb1-secret-volume\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.334575 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl2bx\" (UniqueName: \"kubernetes.io/projected/1a1c1792-e047-413d-8667-791d14563fb1-kube-api-access-tl2bx\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.335620 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a1c1792-e047-413d-8667-791d14563fb1-config-volume\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.345977 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a1c1792-e047-413d-8667-791d14563fb1-secret-volume\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.382889 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl2bx\" (UniqueName: \"kubernetes.io/projected/1a1c1792-e047-413d-8667-791d14563fb1-kube-api-access-tl2bx\") pod \"collect-profiles-29332290-hcp5t\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.466453 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:00 crc kubenswrapper[4945]: I1008 15:30:00.666800 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t"] Oct 08 15:30:00 crc kubenswrapper[4945]: W1008 15:30:00.676128 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a1c1792_e047_413d_8667_791d14563fb1.slice/crio-ed4342d9e7446e8b84fae1d9eb2cf09d13052dde94871983da95b96503d66bd7 WatchSource:0}: Error finding container ed4342d9e7446e8b84fae1d9eb2cf09d13052dde94871983da95b96503d66bd7: Status 404 returned error can't find the container with id ed4342d9e7446e8b84fae1d9eb2cf09d13052dde94871983da95b96503d66bd7 Oct 08 15:30:01 crc kubenswrapper[4945]: I1008 15:30:01.053783 4945 generic.go:334] "Generic (PLEG): container finished" podID="1a1c1792-e047-413d-8667-791d14563fb1" containerID="e2d71abcab67512055d8b9de1a1e952ed6b42795b5b10dcfb45aeee58c5584e3" exitCode=0 Oct 08 15:30:01 crc kubenswrapper[4945]: I1008 15:30:01.053963 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" event={"ID":"1a1c1792-e047-413d-8667-791d14563fb1","Type":"ContainerDied","Data":"e2d71abcab67512055d8b9de1a1e952ed6b42795b5b10dcfb45aeee58c5584e3"} Oct 08 15:30:01 crc kubenswrapper[4945]: I1008 15:30:01.054018 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" event={"ID":"1a1c1792-e047-413d-8667-791d14563fb1","Type":"ContainerStarted","Data":"ed4342d9e7446e8b84fae1d9eb2cf09d13052dde94871983da95b96503d66bd7"} Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.463627 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.580020 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl2bx\" (UniqueName: \"kubernetes.io/projected/1a1c1792-e047-413d-8667-791d14563fb1-kube-api-access-tl2bx\") pod \"1a1c1792-e047-413d-8667-791d14563fb1\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.580222 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a1c1792-e047-413d-8667-791d14563fb1-config-volume\") pod \"1a1c1792-e047-413d-8667-791d14563fb1\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.580275 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a1c1792-e047-413d-8667-791d14563fb1-secret-volume\") pod \"1a1c1792-e047-413d-8667-791d14563fb1\" (UID: \"1a1c1792-e047-413d-8667-791d14563fb1\") " Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.580888 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a1c1792-e047-413d-8667-791d14563fb1-config-volume" (OuterVolumeSpecName: "config-volume") pod "1a1c1792-e047-413d-8667-791d14563fb1" (UID: "1a1c1792-e047-413d-8667-791d14563fb1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.585234 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1c1792-e047-413d-8667-791d14563fb1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1a1c1792-e047-413d-8667-791d14563fb1" (UID: "1a1c1792-e047-413d-8667-791d14563fb1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.586512 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1c1792-e047-413d-8667-791d14563fb1-kube-api-access-tl2bx" (OuterVolumeSpecName: "kube-api-access-tl2bx") pod "1a1c1792-e047-413d-8667-791d14563fb1" (UID: "1a1c1792-e047-413d-8667-791d14563fb1"). InnerVolumeSpecName "kube-api-access-tl2bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.681726 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a1c1792-e047-413d-8667-791d14563fb1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.681771 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl2bx\" (UniqueName: \"kubernetes.io/projected/1a1c1792-e047-413d-8667-791d14563fb1-kube-api-access-tl2bx\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.681787 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a1c1792-e047-413d-8667-791d14563fb1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:03 crc kubenswrapper[4945]: I1008 15:30:03.805143 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:30:04 crc kubenswrapper[4945]: I1008 15:30:04.077457 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" event={"ID":"ef574f9e-15f9-42fe-8593-5575c31fe0d9","Type":"ContainerStarted","Data":"39c2dc2f9e3fd63ead40e26a90b691b81074c21652bb2bd63faa9f1011a4d204"} Oct 08 15:30:04 crc kubenswrapper[4945]: I1008 15:30:04.078863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" event={"ID":"1a1c1792-e047-413d-8667-791d14563fb1","Type":"ContainerDied","Data":"ed4342d9e7446e8b84fae1d9eb2cf09d13052dde94871983da95b96503d66bd7"} Oct 08 15:30:04 crc kubenswrapper[4945]: I1008 15:30:04.078898 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed4342d9e7446e8b84fae1d9eb2cf09d13052dde94871983da95b96503d66bd7" Oct 08 15:30:04 crc kubenswrapper[4945]: I1008 15:30:04.079005 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t" Oct 08 15:30:08 crc kubenswrapper[4945]: I1008 15:30:08.106983 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" event={"ID":"ef574f9e-15f9-42fe-8593-5575c31fe0d9","Type":"ContainerStarted","Data":"d9df25a5beeeb8858a31eb3904df558ac0a9d145a7458ed0d70f46cbe8e74956"} Oct 08 15:30:08 crc kubenswrapper[4945]: I1008 15:30:08.108691 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" Oct 08 15:30:08 crc kubenswrapper[4945]: I1008 15:30:08.112409 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" Oct 08 15:30:08 crc kubenswrapper[4945]: I1008 15:30:08.144554 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-75967d884c-qslk5" podStartSLOduration=2.459204365 podStartE2EDuration="10.144534441s" podCreationTimestamp="2025-10-08 15:29:58 +0000 UTC" firstStartedPulling="2025-10-08 15:29:59.686797901 +0000 UTC m=+1009.040712842" lastFinishedPulling="2025-10-08 15:30:07.372128007 +0000 UTC m=+1016.726042918" observedRunningTime="2025-10-08 15:30:08.144323016 +0000 UTC m=+1017.498237917" watchObservedRunningTime="2025-10-08 15:30:08.144534441 +0000 UTC m=+1017.498449352" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.008220 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm"] Oct 08 15:30:40 crc kubenswrapper[4945]: E1008 15:30:40.009070 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1c1792-e047-413d-8667-791d14563fb1" containerName="collect-profiles" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.009089 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1c1792-e047-413d-8667-791d14563fb1" containerName="collect-profiles" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.009253 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1c1792-e047-413d-8667-791d14563fb1" containerName="collect-profiles" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.010047 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.013464 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-sppx6" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.022728 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.023894 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.026089 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-42wrm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.031938 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.046755 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.047947 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.049900 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hqrn9" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.056126 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.061462 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzs4k\" (UniqueName: \"kubernetes.io/projected/441d050d-f359-46d8-8c29-0b1506c1f240-kube-api-access-jzs4k\") pod \"cinder-operator-controller-manager-59cdc64769-h27l8\" (UID: \"441d050d-f359-46d8-8c29-0b1506c1f240\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.061643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4842j\" (UniqueName: \"kubernetes.io/projected/2747410b-55fa-4a7f-9146-5662ee48cce7-kube-api-access-4842j\") pod \"barbican-operator-controller-manager-64f84fcdbb-7b8pm\" (UID: \"2747410b-55fa-4a7f-9146-5662ee48cce7\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.061671 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w627b\" (UniqueName: \"kubernetes.io/projected/c3ae7aef-2f32-4d2d-b8be-e29dd91ede71-kube-api-access-w627b\") pod \"designate-operator-controller-manager-687df44cdb-4fjrf\" (UID: \"c3ae7aef-2f32-4d2d-b8be-e29dd91ede71\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.079210 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.080533 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.082344 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-9crx8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.083686 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.092746 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.131119 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.132462 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.136480 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-g9n2n" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.162511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzs4k\" (UniqueName: \"kubernetes.io/projected/441d050d-f359-46d8-8c29-0b1506c1f240-kube-api-access-jzs4k\") pod \"cinder-operator-controller-manager-59cdc64769-h27l8\" (UID: \"441d050d-f359-46d8-8c29-0b1506c1f240\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.162628 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4842j\" (UniqueName: \"kubernetes.io/projected/2747410b-55fa-4a7f-9146-5662ee48cce7-kube-api-access-4842j\") pod \"barbican-operator-controller-manager-64f84fcdbb-7b8pm\" (UID: \"2747410b-55fa-4a7f-9146-5662ee48cce7\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.162658 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w627b\" (UniqueName: \"kubernetes.io/projected/c3ae7aef-2f32-4d2d-b8be-e29dd91ede71-kube-api-access-w627b\") pod \"designate-operator-controller-manager-687df44cdb-4fjrf\" (UID: \"c3ae7aef-2f32-4d2d-b8be-e29dd91ede71\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.168191 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.169750 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.171879 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-pbhp2" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.172036 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.173181 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.175049 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.175229 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gch9v" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.187183 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.191529 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.200537 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.204602 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w627b\" (UniqueName: \"kubernetes.io/projected/c3ae7aef-2f32-4d2d-b8be-e29dd91ede71-kube-api-access-w627b\") pod \"designate-operator-controller-manager-687df44cdb-4fjrf\" (UID: \"c3ae7aef-2f32-4d2d-b8be-e29dd91ede71\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.208980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzs4k\" (UniqueName: \"kubernetes.io/projected/441d050d-f359-46d8-8c29-0b1506c1f240-kube-api-access-jzs4k\") pod \"cinder-operator-controller-manager-59cdc64769-h27l8\" (UID: \"441d050d-f359-46d8-8c29-0b1506c1f240\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.215736 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4842j\" (UniqueName: \"kubernetes.io/projected/2747410b-55fa-4a7f-9146-5662ee48cce7-kube-api-access-4842j\") pod \"barbican-operator-controller-manager-64f84fcdbb-7b8pm\" (UID: \"2747410b-55fa-4a7f-9146-5662ee48cce7\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.217239 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.225652 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.225815 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.229319 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.230304 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.236394 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-lnqz9" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.236581 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-pl4p7" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.244375 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.245322 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.247274 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-cqkxt" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.258031 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.258970 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.262047 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2zr58" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.263718 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw6vn\" (UniqueName: \"kubernetes.io/projected/5c314cb3-27af-4a08-b96b-91f9a70f7524-kube-api-access-tw6vn\") pod \"glance-operator-controller-manager-7bb46cd7d-kx5k7\" (UID: \"5c314cb3-27af-4a08-b96b-91f9a70f7524\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.263748 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25b22\" (UniqueName: \"kubernetes.io/projected/84979368-496d-43c1-b792-20c167a2e5d5-kube-api-access-25b22\") pod \"heat-operator-controller-manager-6d9967f8dd-62hpj\" (UID: \"84979368-496d-43c1-b792-20c167a2e5d5\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.282001 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.287217 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.295514 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.304263 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.305705 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.307627 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-mxk8w" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.322900 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.329225 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.349960 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.351172 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.351509 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.353261 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.372259 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-n8s2l" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.390784 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcm2n\" (UniqueName: \"kubernetes.io/projected/f11222c0-9d0f-4d3e-945f-b991135e4235-kube-api-access-pcm2n\") pod \"infra-operator-controller-manager-585fc5b659-r9mp8\" (UID: \"f11222c0-9d0f-4d3e-945f-b991135e4235\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.391012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d57m4\" (UniqueName: \"kubernetes.io/projected/b32201a1-e841-4c4c-8988-0b866243b255-kube-api-access-d57m4\") pod \"horizon-operator-controller-manager-6d74794d9b-j9vft\" (UID: \"b32201a1-e841-4c4c-8988-0b866243b255\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.391140 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn9j4\" (UniqueName: \"kubernetes.io/projected/25cc3669-bdfd-44eb-83d0-85ca303250ea-kube-api-access-wn9j4\") pod \"ironic-operator-controller-manager-74cb5cbc49-hk5rr\" (UID: \"25cc3669-bdfd-44eb-83d0-85ca303250ea\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.391540 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f11222c0-9d0f-4d3e-945f-b991135e4235-cert\") pod \"infra-operator-controller-manager-585fc5b659-r9mp8\" (UID: \"f11222c0-9d0f-4d3e-945f-b991135e4235\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.391758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw6vn\" (UniqueName: \"kubernetes.io/projected/5c314cb3-27af-4a08-b96b-91f9a70f7524-kube-api-access-tw6vn\") pod \"glance-operator-controller-manager-7bb46cd7d-kx5k7\" (UID: \"5c314cb3-27af-4a08-b96b-91f9a70f7524\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.391866 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25b22\" (UniqueName: \"kubernetes.io/projected/84979368-496d-43c1-b792-20c167a2e5d5-kube-api-access-25b22\") pod \"heat-operator-controller-manager-6d9967f8dd-62hpj\" (UID: \"84979368-496d-43c1-b792-20c167a2e5d5\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.392137 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gskk2\" (UniqueName: \"kubernetes.io/projected/cbe3a47b-033a-410d-b2be-bd38e4b9c0ab-kube-api-access-gskk2\") pod \"manila-operator-controller-manager-59578bc799-ht7cm\" (UID: \"cbe3a47b-033a-410d-b2be-bd38e4b9c0ab\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.392317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz2m8\" (UniqueName: \"kubernetes.io/projected/344902f1-cf0c-4390-8f2c-2fe41cf9ccd9-kube-api-access-bz2m8\") pod \"keystone-operator-controller-manager-ddb98f99b-lr6hn\" (UID: \"344902f1-cf0c-4390-8f2c-2fe41cf9ccd9\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.392553 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxtvr\" (UniqueName: \"kubernetes.io/projected/b2ca2442-7e23-40a8-9ca1-1ef11251d34d-kube-api-access-qxtvr\") pod \"mariadb-operator-controller-manager-5777b4f897-p7bfn\" (UID: \"b2ca2442-7e23-40a8-9ca1-1ef11251d34d\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.396197 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.444270 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.445607 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.447250 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw6vn\" (UniqueName: \"kubernetes.io/projected/5c314cb3-27af-4a08-b96b-91f9a70f7524-kube-api-access-tw6vn\") pod \"glance-operator-controller-manager-7bb46cd7d-kx5k7\" (UID: \"5c314cb3-27af-4a08-b96b-91f9a70f7524\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.447968 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-pddrv" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.447972 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25b22\" (UniqueName: \"kubernetes.io/projected/84979368-496d-43c1-b792-20c167a2e5d5-kube-api-access-25b22\") pod \"heat-operator-controller-manager-6d9967f8dd-62hpj\" (UID: \"84979368-496d-43c1-b792-20c167a2e5d5\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.453574 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.455051 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.468536 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.469969 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.475974 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.477482 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.478704 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nsnnw" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.481257 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.481453 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-cpdkk" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.481550 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.489915 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.491588 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-c8h57"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.492883 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496085 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz2m8\" (UniqueName: \"kubernetes.io/projected/344902f1-cf0c-4390-8f2c-2fe41cf9ccd9-kube-api-access-bz2m8\") pod \"keystone-operator-controller-manager-ddb98f99b-lr6hn\" (UID: \"344902f1-cf0c-4390-8f2c-2fe41cf9ccd9\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496330 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxtvr\" (UniqueName: \"kubernetes.io/projected/b2ca2442-7e23-40a8-9ca1-1ef11251d34d-kube-api-access-qxtvr\") pod \"mariadb-operator-controller-manager-5777b4f897-p7bfn\" (UID: \"b2ca2442-7e23-40a8-9ca1-1ef11251d34d\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gd7j\" (UniqueName: \"kubernetes.io/projected/1112a081-2d10-48e2-b345-8ef841c5dc50-kube-api-access-2gd7j\") pod \"neutron-operator-controller-manager-797d478b46-t56f2\" (UID: \"1112a081-2d10-48e2-b345-8ef841c5dc50\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496410 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcm2n\" (UniqueName: \"kubernetes.io/projected/f11222c0-9d0f-4d3e-945f-b991135e4235-kube-api-access-pcm2n\") pod \"infra-operator-controller-manager-585fc5b659-r9mp8\" (UID: \"f11222c0-9d0f-4d3e-945f-b991135e4235\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496428 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d57m4\" (UniqueName: \"kubernetes.io/projected/b32201a1-e841-4c4c-8988-0b866243b255-kube-api-access-d57m4\") pod \"horizon-operator-controller-manager-6d74794d9b-j9vft\" (UID: \"b32201a1-e841-4c4c-8988-0b866243b255\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn9j4\" (UniqueName: \"kubernetes.io/projected/25cc3669-bdfd-44eb-83d0-85ca303250ea-kube-api-access-wn9j4\") pod \"ironic-operator-controller-manager-74cb5cbc49-hk5rr\" (UID: \"25cc3669-bdfd-44eb-83d0-85ca303250ea\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496471 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f11222c0-9d0f-4d3e-945f-b991135e4235-cert\") pod \"infra-operator-controller-manager-585fc5b659-r9mp8\" (UID: \"f11222c0-9d0f-4d3e-945f-b991135e4235\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496493 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gskk2\" (UniqueName: \"kubernetes.io/projected/cbe3a47b-033a-410d-b2be-bd38e4b9c0ab-kube-api-access-gskk2\") pod \"manila-operator-controller-manager-59578bc799-ht7cm\" (UID: \"cbe3a47b-033a-410d-b2be-bd38e4b9c0ab\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.496521 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2g44\" (UniqueName: \"kubernetes.io/projected/3501f2c3-1b0d-4f7a-9fe2-2516f0088993-kube-api-access-p2g44\") pod \"nova-operator-controller-manager-57bb74c7bf-lwppj\" (UID: \"3501f2c3-1b0d-4f7a-9fe2-2516f0088993\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.504047 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f11222c0-9d0f-4d3e-945f-b991135e4235-cert\") pod \"infra-operator-controller-manager-585fc5b659-r9mp8\" (UID: \"f11222c0-9d0f-4d3e-945f-b991135e4235\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.504685 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-59p2t" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.505982 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-c8h57"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.521672 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.523166 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.529568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn9j4\" (UniqueName: \"kubernetes.io/projected/25cc3669-bdfd-44eb-83d0-85ca303250ea-kube-api-access-wn9j4\") pod \"ironic-operator-controller-manager-74cb5cbc49-hk5rr\" (UID: \"25cc3669-bdfd-44eb-83d0-85ca303250ea\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.531031 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x47kd" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.532777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxtvr\" (UniqueName: \"kubernetes.io/projected/b2ca2442-7e23-40a8-9ca1-1ef11251d34d-kube-api-access-qxtvr\") pod \"mariadb-operator-controller-manager-5777b4f897-p7bfn\" (UID: \"b2ca2442-7e23-40a8-9ca1-1ef11251d34d\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.535949 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gskk2\" (UniqueName: \"kubernetes.io/projected/cbe3a47b-033a-410d-b2be-bd38e4b9c0ab-kube-api-access-gskk2\") pod \"manila-operator-controller-manager-59578bc799-ht7cm\" (UID: \"cbe3a47b-033a-410d-b2be-bd38e4b9c0ab\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.536977 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.558544 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d57m4\" (UniqueName: \"kubernetes.io/projected/b32201a1-e841-4c4c-8988-0b866243b255-kube-api-access-d57m4\") pod \"horizon-operator-controller-manager-6d74794d9b-j9vft\" (UID: \"b32201a1-e841-4c4c-8988-0b866243b255\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.570229 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.571545 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.572807 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcm2n\" (UniqueName: \"kubernetes.io/projected/f11222c0-9d0f-4d3e-945f-b991135e4235-kube-api-access-pcm2n\") pod \"infra-operator-controller-manager-585fc5b659-r9mp8\" (UID: \"f11222c0-9d0f-4d3e-945f-b991135e4235\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.573463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.575852 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-vsppv" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.579649 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.580247 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz2m8\" (UniqueName: \"kubernetes.io/projected/344902f1-cf0c-4390-8f2c-2fe41cf9ccd9-kube-api-access-bz2m8\") pod \"keystone-operator-controller-manager-ddb98f99b-lr6hn\" (UID: \"344902f1-cf0c-4390-8f2c-2fe41cf9ccd9\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.583448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.599291 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57gsk\" (UniqueName: \"kubernetes.io/projected/efd39ce1-ac6e-449d-b32f-fc451ffde97e-kube-api-access-57gsk\") pod \"placement-operator-controller-manager-664664cb68-c8h57\" (UID: \"efd39ce1-ac6e-449d-b32f-fc451ffde97e\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.599333 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.599372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvnbl\" (UniqueName: \"kubernetes.io/projected/9555aa6e-7b8f-4354-96e8-e639719e1b56-kube-api-access-qvnbl\") pod \"octavia-operator-controller-manager-6d7c7ddf95-tb4rv\" (UID: \"9555aa6e-7b8f-4354-96e8-e639719e1b56\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.599402 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2g44\" (UniqueName: \"kubernetes.io/projected/3501f2c3-1b0d-4f7a-9fe2-2516f0088993-kube-api-access-p2g44\") pod \"nova-operator-controller-manager-57bb74c7bf-lwppj\" (UID: \"3501f2c3-1b0d-4f7a-9fe2-2516f0088993\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.599430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r54d\" (UniqueName: \"kubernetes.io/projected/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-kube-api-access-4r54d\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.599527 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fxlt\" (UniqueName: \"kubernetes.io/projected/400ef757-79e4-4dd6-8c0a-2440835ac048-kube-api-access-5fxlt\") pod \"ovn-operator-controller-manager-6f96f8c84-5t8fx\" (UID: \"400ef757-79e4-4dd6-8c0a-2440835ac048\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.599609 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gd7j\" (UniqueName: \"kubernetes.io/projected/1112a081-2d10-48e2-b345-8ef841c5dc50-kube-api-access-2gd7j\") pod \"neutron-operator-controller-manager-797d478b46-t56f2\" (UID: \"1112a081-2d10-48e2-b345-8ef841c5dc50\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.608840 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.618477 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.638700 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2g44\" (UniqueName: \"kubernetes.io/projected/3501f2c3-1b0d-4f7a-9fe2-2516f0088993-kube-api-access-p2g44\") pod \"nova-operator-controller-manager-57bb74c7bf-lwppj\" (UID: \"3501f2c3-1b0d-4f7a-9fe2-2516f0088993\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.640606 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.640979 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gd7j\" (UniqueName: \"kubernetes.io/projected/1112a081-2d10-48e2-b345-8ef841c5dc50-kube-api-access-2gd7j\") pod \"neutron-operator-controller-manager-797d478b46-t56f2\" (UID: \"1112a081-2d10-48e2-b345-8ef841c5dc50\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.642033 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.645793 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-x47p5" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.670675 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.696743 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-df6684d58-22z48"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.698629 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhndz\" (UniqueName: \"kubernetes.io/projected/f8b50124-bf7a-497b-bdc6-42bbbe564a7d-kube-api-access-qhndz\") pod \"telemetry-operator-controller-manager-775776c574-rn55t\" (UID: \"f8b50124-bf7a-497b-bdc6-42bbbe564a7d\") " pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701356 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvnbl\" (UniqueName: \"kubernetes.io/projected/9555aa6e-7b8f-4354-96e8-e639719e1b56-kube-api-access-qvnbl\") pod \"octavia-operator-controller-manager-6d7c7ddf95-tb4rv\" (UID: \"9555aa6e-7b8f-4354-96e8-e639719e1b56\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r54d\" (UniqueName: \"kubernetes.io/projected/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-kube-api-access-4r54d\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fxlt\" (UniqueName: \"kubernetes.io/projected/400ef757-79e4-4dd6-8c0a-2440835ac048-kube-api-access-5fxlt\") pod \"ovn-operator-controller-manager-6f96f8c84-5t8fx\" (UID: \"400ef757-79e4-4dd6-8c0a-2440835ac048\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701690 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w44k\" (UniqueName: \"kubernetes.io/projected/43c420bf-cf21-4caa-90ce-0a9009d3bc45-kube-api-access-5w44k\") pod \"swift-operator-controller-manager-5f4d5dfdc6-qxbqr\" (UID: \"43c420bf-cf21-4caa-90ce-0a9009d3bc45\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57gsk\" (UniqueName: \"kubernetes.io/projected/efd39ce1-ac6e-449d-b32f-fc451ffde97e-kube-api-access-57gsk\") pod \"placement-operator-controller-manager-664664cb68-c8h57\" (UID: \"efd39ce1-ac6e-449d-b32f-fc451ffde97e\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701569 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-nlcgm" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.701944 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:40 crc kubenswrapper[4945]: E1008 15:30:40.702135 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 15:30:40 crc kubenswrapper[4945]: E1008 15:30:40.702231 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert podName:a8d00f8c-42c6-4d99-bb76-bdd491597e3a nodeName:}" failed. No retries permitted until 2025-10-08 15:30:41.202215198 +0000 UTC m=+1050.556130099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" (UID: "a8d00f8c-42c6-4d99-bb76-bdd491597e3a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.718220 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.724739 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r54d\" (UniqueName: \"kubernetes.io/projected/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-kube-api-access-4r54d\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.729211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvnbl\" (UniqueName: \"kubernetes.io/projected/9555aa6e-7b8f-4354-96e8-e639719e1b56-kube-api-access-qvnbl\") pod \"octavia-operator-controller-manager-6d7c7ddf95-tb4rv\" (UID: \"9555aa6e-7b8f-4354-96e8-e639719e1b56\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.730033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fxlt\" (UniqueName: \"kubernetes.io/projected/400ef757-79e4-4dd6-8c0a-2440835ac048-kube-api-access-5fxlt\") pod \"ovn-operator-controller-manager-6f96f8c84-5t8fx\" (UID: \"400ef757-79e4-4dd6-8c0a-2440835ac048\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.730086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57gsk\" (UniqueName: \"kubernetes.io/projected/efd39ce1-ac6e-449d-b32f-fc451ffde97e-kube-api-access-57gsk\") pod \"placement-operator-controller-manager-664664cb68-c8h57\" (UID: \"efd39ce1-ac6e-449d-b32f-fc451ffde97e\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.730910 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-df6684d58-22z48"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.751713 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.766831 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.768027 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.774360 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.776003 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.778272 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-fl69d" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.788427 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.792054 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.803313 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w44k\" (UniqueName: \"kubernetes.io/projected/43c420bf-cf21-4caa-90ce-0a9009d3bc45-kube-api-access-5w44k\") pod \"swift-operator-controller-manager-5f4d5dfdc6-qxbqr\" (UID: \"43c420bf-cf21-4caa-90ce-0a9009d3bc45\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.804552 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmmmk\" (UniqueName: \"kubernetes.io/projected/e0437b00-8fcc-412a-90fa-32ea56c71421-kube-api-access-qmmmk\") pod \"watcher-operator-controller-manager-df6684d58-22z48\" (UID: \"e0437b00-8fcc-412a-90fa-32ea56c71421\") " pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.804702 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mc2k\" (UniqueName: \"kubernetes.io/projected/28b9d7cf-ad46-4c2a-a942-f7b20c054216-kube-api-access-7mc2k\") pod \"test-operator-controller-manager-74665f6cdc-dwfj9\" (UID: \"28b9d7cf-ad46-4c2a-a942-f7b20c054216\") " pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.804742 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhndz\" (UniqueName: \"kubernetes.io/projected/f8b50124-bf7a-497b-bdc6-42bbbe564a7d-kube-api-access-qhndz\") pod \"telemetry-operator-controller-manager-775776c574-rn55t\" (UID: \"f8b50124-bf7a-497b-bdc6-42bbbe564a7d\") " pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.807702 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.833357 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.834597 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.837666 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.840668 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-tgk72" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.841053 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhndz\" (UniqueName: \"kubernetes.io/projected/f8b50124-bf7a-497b-bdc6-42bbbe564a7d-kube-api-access-qhndz\") pod \"telemetry-operator-controller-manager-775776c574-rn55t\" (UID: \"f8b50124-bf7a-497b-bdc6-42bbbe564a7d\") " pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.843148 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w44k\" (UniqueName: \"kubernetes.io/projected/43c420bf-cf21-4caa-90ce-0a9009d3bc45-kube-api-access-5w44k\") pod \"swift-operator-controller-manager-5f4d5dfdc6-qxbqr\" (UID: \"43c420bf-cf21-4caa-90ce-0a9009d3bc45\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.849789 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx"] Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.853161 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.858259 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.908833 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c42bd8a4-d1ed-446c-9295-03cfe33588de-cert\") pod \"openstack-operator-controller-manager-56688bc7d6-h2wfx\" (UID: \"c42bd8a4-d1ed-446c-9295-03cfe33588de\") " pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.909087 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps8bk\" (UniqueName: \"kubernetes.io/projected/c42bd8a4-d1ed-446c-9295-03cfe33588de-kube-api-access-ps8bk\") pod \"openstack-operator-controller-manager-56688bc7d6-h2wfx\" (UID: \"c42bd8a4-d1ed-446c-9295-03cfe33588de\") " pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.909254 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmmmk\" (UniqueName: \"kubernetes.io/projected/e0437b00-8fcc-412a-90fa-32ea56c71421-kube-api-access-qmmmk\") pod \"watcher-operator-controller-manager-df6684d58-22z48\" (UID: \"e0437b00-8fcc-412a-90fa-32ea56c71421\") " pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.909357 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mc2k\" (UniqueName: \"kubernetes.io/projected/28b9d7cf-ad46-4c2a-a942-f7b20c054216-kube-api-access-7mc2k\") pod \"test-operator-controller-manager-74665f6cdc-dwfj9\" (UID: \"28b9d7cf-ad46-4c2a-a942-f7b20c054216\") " pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.922531 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.956810 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmmmk\" (UniqueName: \"kubernetes.io/projected/e0437b00-8fcc-412a-90fa-32ea56c71421-kube-api-access-qmmmk\") pod \"watcher-operator-controller-manager-df6684d58-22z48\" (UID: \"e0437b00-8fcc-412a-90fa-32ea56c71421\") " pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.958154 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" Oct 08 15:30:40 crc kubenswrapper[4945]: I1008 15:30:40.964680 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mc2k\" (UniqueName: \"kubernetes.io/projected/28b9d7cf-ad46-4c2a-a942-f7b20c054216-kube-api-access-7mc2k\") pod \"test-operator-controller-manager-74665f6cdc-dwfj9\" (UID: \"28b9d7cf-ad46-4c2a-a942-f7b20c054216\") " pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.005807 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.014523 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c42bd8a4-d1ed-446c-9295-03cfe33588de-cert\") pod \"openstack-operator-controller-manager-56688bc7d6-h2wfx\" (UID: \"c42bd8a4-d1ed-446c-9295-03cfe33588de\") " pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.015693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps8bk\" (UniqueName: \"kubernetes.io/projected/c42bd8a4-d1ed-446c-9295-03cfe33588de-kube-api-access-ps8bk\") pod \"openstack-operator-controller-manager-56688bc7d6-h2wfx\" (UID: \"c42bd8a4-d1ed-446c-9295-03cfe33588de\") " pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.015820 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw4l2\" (UniqueName: \"kubernetes.io/projected/1d068bb6-a523-4903-95b7-7c62862c2bb3-kube-api-access-gw4l2\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx\" (UID: \"1d068bb6-a523-4903-95b7-7c62862c2bb3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" Oct 08 15:30:41 crc kubenswrapper[4945]: E1008 15:30:41.015290 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 08 15:30:41 crc kubenswrapper[4945]: E1008 15:30:41.016153 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c42bd8a4-d1ed-446c-9295-03cfe33588de-cert podName:c42bd8a4-d1ed-446c-9295-03cfe33588de nodeName:}" failed. No retries permitted until 2025-10-08 15:30:41.51612988 +0000 UTC m=+1050.870044781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c42bd8a4-d1ed-446c-9295-03cfe33588de-cert") pod "openstack-operator-controller-manager-56688bc7d6-h2wfx" (UID: "c42bd8a4-d1ed-446c-9295-03cfe33588de") : secret "webhook-server-cert" not found Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.057598 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps8bk\" (UniqueName: \"kubernetes.io/projected/c42bd8a4-d1ed-446c-9295-03cfe33588de-kube-api-access-ps8bk\") pod \"openstack-operator-controller-manager-56688bc7d6-h2wfx\" (UID: \"c42bd8a4-d1ed-446c-9295-03cfe33588de\") " pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.120510 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw4l2\" (UniqueName: \"kubernetes.io/projected/1d068bb6-a523-4903-95b7-7c62862c2bb3-kube-api-access-gw4l2\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx\" (UID: \"1d068bb6-a523-4903-95b7-7c62862c2bb3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.135676 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.147089 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw4l2\" (UniqueName: \"kubernetes.io/projected/1d068bb6-a523-4903-95b7-7c62862c2bb3-kube-api-access-gw4l2\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx\" (UID: \"1d068bb6-a523-4903-95b7-7c62862c2bb3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.222297 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:41 crc kubenswrapper[4945]: E1008 15:30:41.222915 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 15:30:41 crc kubenswrapper[4945]: E1008 15:30:41.222951 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert podName:a8d00f8c-42c6-4d99-bb76-bdd491597e3a nodeName:}" failed. No retries permitted until 2025-10-08 15:30:42.222938164 +0000 UTC m=+1051.576853065 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" (UID: "a8d00f8c-42c6-4d99-bb76-bdd491597e3a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.391147 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.478480 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8"] Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.527334 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c42bd8a4-d1ed-446c-9295-03cfe33588de-cert\") pod \"openstack-operator-controller-manager-56688bc7d6-h2wfx\" (UID: \"c42bd8a4-d1ed-446c-9295-03cfe33588de\") " pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.531186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c42bd8a4-d1ed-446c-9295-03cfe33588de-cert\") pod \"openstack-operator-controller-manager-56688bc7d6-h2wfx\" (UID: \"c42bd8a4-d1ed-446c-9295-03cfe33588de\") " pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.648187 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.692670 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj"] Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.728631 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf"] Oct 08 15:30:41 crc kubenswrapper[4945]: I1008 15:30:41.743643 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm"] Oct 08 15:30:41 crc kubenswrapper[4945]: W1008 15:30:41.750386 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3ae7aef_2f32_4d2d_b8be_e29dd91ede71.slice/crio-2f04dc877af5be4cbb48dd7799ee4dfbe6a03a20a572dcf1b478345126c7d342 WatchSource:0}: Error finding container 2f04dc877af5be4cbb48dd7799ee4dfbe6a03a20a572dcf1b478345126c7d342: Status 404 returned error can't find the container with id 2f04dc877af5be4cbb48dd7799ee4dfbe6a03a20a572dcf1b478345126c7d342 Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.158352 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.171349 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv"] Oct 08 15:30:42 crc kubenswrapper[4945]: W1008 15:30:42.182857 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9555aa6e_7b8f_4354_96e8_e639719e1b56.slice/crio-c70b7c764ffd7fd67814a25ecbfe1d5c5b3817c011000bd5fe237b43ebae163a WatchSource:0}: Error finding container c70b7c764ffd7fd67814a25ecbfe1d5c5b3817c011000bd5fe237b43ebae163a: Status 404 returned error can't find the container with id c70b7c764ffd7fd67814a25ecbfe1d5c5b3817c011000bd5fe237b43ebae163a Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.217052 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.232353 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.236869 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.239778 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.251086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8d00f8c-42c6-4d99-bb76-bdd491597e3a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt\" (UID: \"a8d00f8c-42c6-4d99-bb76-bdd491597e3a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.297841 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.322742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.324178 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.332552 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.357172 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.376240 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-c8h57"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.380176 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.388166 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.388666 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.394220 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.396803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" event={"ID":"344902f1-cf0c-4390-8f2c-2fe41cf9ccd9","Type":"ContainerStarted","Data":"9904ff6a0914fe4d32c55929da166648752cc60599f677063ba2189ddf6e2aa8"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.398463 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.401292 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" event={"ID":"3501f2c3-1b0d-4f7a-9fe2-2516f0088993","Type":"ContainerStarted","Data":"647dc9b19055280199e82cc5d9dac534772cdc6e5d921ff2c43ef8c70cb94c75"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.408485 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-df6684d58-22z48"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.424687 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" event={"ID":"9555aa6e-7b8f-4354-96e8-e639719e1b56","Type":"ContainerStarted","Data":"c70b7c764ffd7fd67814a25ecbfe1d5c5b3817c011000bd5fe237b43ebae163a"} Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.424911 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-57gsk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-c8h57_openstack-operators(efd39ce1-ac6e-449d-b32f-fc451ffde97e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.425014 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pcm2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-585fc5b659-r9mp8_openstack-operators(f11222c0-9d0f-4d3e-945f-b991135e4235): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.429262 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx"] Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.440947 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" event={"ID":"5c314cb3-27af-4a08-b96b-91f9a70f7524","Type":"ContainerStarted","Data":"42510210877c487a60881d38a1630b0e7a56592f04a98704656c91866bf4209e"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.451586 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx"] Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.465407 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7mc2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-74665f6cdc-dwfj9_openstack-operators(28b9d7cf-ad46-4c2a-a942-f7b20c054216): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.484904 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.163:5001/openstack-k8s-operators/watcher-operator:1c247f8dbfb12d23c6a704edea00ffb8713ded2a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qmmmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-df6684d58-22z48_openstack-operators(e0437b00-8fcc-412a-90fa-32ea56c71421): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.485095 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gskk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-59578bc799-ht7cm_openstack-operators(cbe3a47b-033a-410d-b2be-bd38e4b9c0ab): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.500588 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" event={"ID":"2747410b-55fa-4a7f-9146-5662ee48cce7","Type":"ContainerStarted","Data":"6512e041868fe75bd81d05511b026e564c50179d0da9ed63e0c462013587a1fa"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.515408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" event={"ID":"25cc3669-bdfd-44eb-83d0-85ca303250ea","Type":"ContainerStarted","Data":"5039b15eb72662eee8b16f5455b2028b8779642174d2de2cdfa6ccb33e92284c"} Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.519125 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gw4l2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx_openstack-operators(1d068bb6-a523-4903-95b7-7c62862c2bb3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.520584 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" podUID="1d068bb6-a523-4903-95b7-7c62862c2bb3" Oct 08 15:30:42 crc kubenswrapper[4945]: E1008 15:30:42.524202 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gd7j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-797d478b46-t56f2_openstack-operators(1112a081-2d10-48e2-b345-8ef841c5dc50): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.542416 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" event={"ID":"c3ae7aef-2f32-4d2d-b8be-e29dd91ede71","Type":"ContainerStarted","Data":"2f04dc877af5be4cbb48dd7799ee4dfbe6a03a20a572dcf1b478345126c7d342"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.603220 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" event={"ID":"84979368-496d-43c1-b792-20c167a2e5d5","Type":"ContainerStarted","Data":"4643b02068ba7059cbe45de45a4622966054cb5ab28611a83239a74e1c64d632"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.608162 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" event={"ID":"441d050d-f359-46d8-8c29-0b1506c1f240","Type":"ContainerStarted","Data":"547c9463d08a9be44475a8884ff3bebe413fbd279333a341fe382f97d2f7f14c"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.624673 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" event={"ID":"b32201a1-e841-4c4c-8988-0b866243b255","Type":"ContainerStarted","Data":"df8e5bdb8d992c64431a85a72e1ab41763144aa7266fa75f47df3c3f1b98a204"} Oct 08 15:30:42 crc kubenswrapper[4945]: I1008 15:30:42.627457 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" event={"ID":"400ef757-79e4-4dd6-8c0a-2440835ac048","Type":"ContainerStarted","Data":"c73ccad075de85e1035563ac972a5f9dccf39de21e4a3dcbccc8fad32f22a156"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.045565 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt"] Oct 08 15:30:43 crc kubenswrapper[4945]: W1008 15:30:43.063972 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8d00f8c_42c6_4d99_bb76_bdd491597e3a.slice/crio-b7bbca17db580d92f6bfe5a0099cd19b59735a1faedaeaf4b871eb6f7d52b0fb WatchSource:0}: Error finding container b7bbca17db580d92f6bfe5a0099cd19b59735a1faedaeaf4b871eb6f7d52b0fb: Status 404 returned error can't find the container with id b7bbca17db580d92f6bfe5a0099cd19b59735a1faedaeaf4b871eb6f7d52b0fb Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.070640 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" podUID="28b9d7cf-ad46-4c2a-a942-f7b20c054216" Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.072216 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" podUID="cbe3a47b-033a-410d-b2be-bd38e4b9c0ab" Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.074293 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" podUID="efd39ce1-ac6e-449d-b32f-fc451ffde97e" Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.142169 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" podUID="e0437b00-8fcc-412a-90fa-32ea56c71421" Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.142684 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" podUID="1112a081-2d10-48e2-b345-8ef841c5dc50" Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.146229 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" podUID="f11222c0-9d0f-4d3e-945f-b991135e4235" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.663009 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" event={"ID":"c42bd8a4-d1ed-446c-9295-03cfe33588de","Type":"ContainerStarted","Data":"31eabcd0497b12828d4097850706c99a5940edaf496d40bfe9cfd021767e01ea"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.663049 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" event={"ID":"c42bd8a4-d1ed-446c-9295-03cfe33588de","Type":"ContainerStarted","Data":"68330bea7e4ebe49ecf83e5fa1da9db3b0c2306423f539cfca515cb435faaa02"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.663060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" event={"ID":"c42bd8a4-d1ed-446c-9295-03cfe33588de","Type":"ContainerStarted","Data":"e7d23357cdff2d45f04fe2a56f60e319e7588aa7fe79eea87d283172ebca0623"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.663217 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.686532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" event={"ID":"b2ca2442-7e23-40a8-9ca1-1ef11251d34d","Type":"ContainerStarted","Data":"24776b0a722c4cf49d5edcbe5a2b839e9d4f447b218168de74c551cde293f1ac"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.695726 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" event={"ID":"1d068bb6-a523-4903-95b7-7c62862c2bb3","Type":"ContainerStarted","Data":"cbe3e58960f0ecd9d57d25a6079dfe60e2c7f82409e37251e65312f780e37d50"} Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.696783 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" podUID="1d068bb6-a523-4903-95b7-7c62862c2bb3" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.698027 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" event={"ID":"1112a081-2d10-48e2-b345-8ef841c5dc50","Type":"ContainerStarted","Data":"d75147fd568979139b9547ed28b0f9562e76c25f4d64b60705dd89d72a407f8e"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.698052 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" event={"ID":"1112a081-2d10-48e2-b345-8ef841c5dc50","Type":"ContainerStarted","Data":"bb77a506e5d5eea80ea832f31f16be1cc3bf179695032ba1654840136cf5d622"} Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.699489 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" podUID="1112a081-2d10-48e2-b345-8ef841c5dc50" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.700139 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" event={"ID":"e0437b00-8fcc-412a-90fa-32ea56c71421","Type":"ContainerStarted","Data":"a8db731f395a01850e4a5d4cdf673b1bffa92fef7d240b9a7ed7645c6ed7c3fd"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.700192 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" event={"ID":"e0437b00-8fcc-412a-90fa-32ea56c71421","Type":"ContainerStarted","Data":"c13d501df06860e758a1392f62154d9ad1bf1d9855150e8e37fa8a090bcbd6aa"} Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.708377 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/openstack-k8s-operators/watcher-operator:1c247f8dbfb12d23c6a704edea00ffb8713ded2a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" podUID="e0437b00-8fcc-412a-90fa-32ea56c71421" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.709898 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" event={"ID":"cbe3a47b-033a-410d-b2be-bd38e4b9c0ab","Type":"ContainerStarted","Data":"460bec22adaaae6ae944005220d4114e50062a8c4caedee224a629f6d317cb43"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.709927 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" event={"ID":"cbe3a47b-033a-410d-b2be-bd38e4b9c0ab","Type":"ContainerStarted","Data":"11f82cb6af25c0cd98744a4d8d2e59b3bc8c82ed17c259cfb69a895d135b15bd"} Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.714845 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" podUID="cbe3a47b-033a-410d-b2be-bd38e4b9c0ab" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.752047 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" event={"ID":"28b9d7cf-ad46-4c2a-a942-f7b20c054216","Type":"ContainerStarted","Data":"b8bc9210c2c29d817d0e9e0eb02abfc7e6a4703113d7c34fdb7ad7d3dee27af9"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.752092 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" event={"ID":"28b9d7cf-ad46-4c2a-a942-f7b20c054216","Type":"ContainerStarted","Data":"e4b1d338debefba4a9cf3405387827872112aa2d2b37db9d1e5376d6fb2f3d84"} Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.756356 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9\\\"\"" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" podUID="28b9d7cf-ad46-4c2a-a942-f7b20c054216" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.757579 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" podStartSLOduration=3.757562989 podStartE2EDuration="3.757562989s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:30:43.735489541 +0000 UTC m=+1053.089404442" watchObservedRunningTime="2025-10-08 15:30:43.757562989 +0000 UTC m=+1053.111477890" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.770519 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" event={"ID":"f11222c0-9d0f-4d3e-945f-b991135e4235","Type":"ContainerStarted","Data":"a9db57a0434f2b4879aae70786b181e169393f4fe8d9d8bf741e9550b92b2f5c"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.770560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" event={"ID":"f11222c0-9d0f-4d3e-945f-b991135e4235","Type":"ContainerStarted","Data":"44a5d34dbb6ca267c46a07a697e72db1696b0190dc3a77235fa65e548faeda0c"} Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.779468 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" podUID="f11222c0-9d0f-4d3e-945f-b991135e4235" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.785298 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" event={"ID":"f8b50124-bf7a-497b-bdc6-42bbbe564a7d","Type":"ContainerStarted","Data":"ad82d875e99b3927651320ae1133ec4934fc4ca81d639d901ad723aa0e347ba2"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.787361 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" event={"ID":"efd39ce1-ac6e-449d-b32f-fc451ffde97e","Type":"ContainerStarted","Data":"0c59bb8222756c7a8f93cbe430272c207e45c396ee8b9ca5563a552add62fd6f"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.787387 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" event={"ID":"efd39ce1-ac6e-449d-b32f-fc451ffde97e","Type":"ContainerStarted","Data":"9086edae9f64cdf7bdedac2f14be0ccf86e2d790c7edb584198520d50d697fe1"} Oct 08 15:30:43 crc kubenswrapper[4945]: E1008 15:30:43.788616 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" podUID="efd39ce1-ac6e-449d-b32f-fc451ffde97e" Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.789081 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" event={"ID":"43c420bf-cf21-4caa-90ce-0a9009d3bc45","Type":"ContainerStarted","Data":"b998ff7dc9eeb7c61cce86228299c344ee8071064fe1bd0c40557ff349f64240"} Oct 08 15:30:43 crc kubenswrapper[4945]: I1008 15:30:43.802340 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" event={"ID":"a8d00f8c-42c6-4d99-bb76-bdd491597e3a","Type":"ContainerStarted","Data":"b7bbca17db580d92f6bfe5a0099cd19b59735a1faedaeaf4b871eb6f7d52b0fb"} Oct 08 15:30:44 crc kubenswrapper[4945]: E1008 15:30:44.811449 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/openstack-k8s-operators/watcher-operator:1c247f8dbfb12d23c6a704edea00ffb8713ded2a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" podUID="e0437b00-8fcc-412a-90fa-32ea56c71421" Oct 08 15:30:44 crc kubenswrapper[4945]: E1008 15:30:44.811466 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" podUID="f11222c0-9d0f-4d3e-945f-b991135e4235" Oct 08 15:30:44 crc kubenswrapper[4945]: E1008 15:30:44.811675 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" podUID="efd39ce1-ac6e-449d-b32f-fc451ffde97e" Oct 08 15:30:44 crc kubenswrapper[4945]: E1008 15:30:44.812244 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" podUID="1112a081-2d10-48e2-b345-8ef841c5dc50" Oct 08 15:30:44 crc kubenswrapper[4945]: E1008 15:30:44.812486 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9\\\"\"" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" podUID="28b9d7cf-ad46-4c2a-a942-f7b20c054216" Oct 08 15:30:44 crc kubenswrapper[4945]: E1008 15:30:44.813375 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" podUID="1d068bb6-a523-4903-95b7-7c62862c2bb3" Oct 08 15:30:44 crc kubenswrapper[4945]: E1008 15:30:44.815878 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" podUID="cbe3a47b-033a-410d-b2be-bd38e4b9c0ab" Oct 08 15:30:51 crc kubenswrapper[4945]: I1008 15:30:51.655969 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-56688bc7d6-h2wfx" Oct 08 15:30:55 crc kubenswrapper[4945]: I1008 15:30:55.901709 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" event={"ID":"c3ae7aef-2f32-4d2d-b8be-e29dd91ede71","Type":"ContainerStarted","Data":"ec28de422133de56bd4f7f3ac20a9af5857f0f6a089088e3942f0e6c46056c22"} Oct 08 15:30:55 crc kubenswrapper[4945]: I1008 15:30:55.904214 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" event={"ID":"84979368-496d-43c1-b792-20c167a2e5d5","Type":"ContainerStarted","Data":"b30c344efb9a8cfe875dd5dee983852c60082e22d95ffc7c3cfd60fe4ebce5c2"} Oct 08 15:30:55 crc kubenswrapper[4945]: I1008 15:30:55.916058 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" event={"ID":"441d050d-f359-46d8-8c29-0b1506c1f240","Type":"ContainerStarted","Data":"603d81581cc6659ec1c38d8ef0061d48bc539ba0e3e594b6d0a21d57cc5a9acb"} Oct 08 15:30:55 crc kubenswrapper[4945]: I1008 15:30:55.928346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" event={"ID":"9555aa6e-7b8f-4354-96e8-e639719e1b56","Type":"ContainerStarted","Data":"6ca442daea5d5af56800df1761a79eeb12c45133090ce878ccfa37dae732b99f"} Oct 08 15:30:55 crc kubenswrapper[4945]: I1008 15:30:55.962559 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" event={"ID":"f8b50124-bf7a-497b-bdc6-42bbbe564a7d","Type":"ContainerStarted","Data":"ff9108552ca3953e706c788ea068411e89196d4451fe7d4ceae3b24bf0c395e1"} Oct 08 15:30:55 crc kubenswrapper[4945]: I1008 15:30:55.978087 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" event={"ID":"400ef757-79e4-4dd6-8c0a-2440835ac048","Type":"ContainerStarted","Data":"6ecfd44b283760cb94738de3dd9b33e3e25aa309b2bcdeb8b40ad01fe7744b8c"} Oct 08 15:30:55 crc kubenswrapper[4945]: I1008 15:30:55.986983 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" event={"ID":"3501f2c3-1b0d-4f7a-9fe2-2516f0088993","Type":"ContainerStarted","Data":"3eba5b64b674283502c336e248d15ece354987c4cd5809a708a7edacc4f9b886"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.020249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" event={"ID":"344902f1-cf0c-4390-8f2c-2fe41cf9ccd9","Type":"ContainerStarted","Data":"616fb5c87ffdc9d40a206cf40cf5b779489e6be5d032f5fd1dfe9f960361fe24"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.025492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" event={"ID":"3501f2c3-1b0d-4f7a-9fe2-2516f0088993","Type":"ContainerStarted","Data":"0156db8f057f46c5d016d353b995211808196bbe8cca07786312dbfac4f265d7"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.026198 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.037186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" event={"ID":"84979368-496d-43c1-b792-20c167a2e5d5","Type":"ContainerStarted","Data":"ae0138cd08454ca0305e3fa488fced628730310bae3368cf299bde38a9a3f5ea"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.037841 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.039334 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" event={"ID":"a8d00f8c-42c6-4d99-bb76-bdd491597e3a","Type":"ContainerStarted","Data":"588056533b627bf79b58a9f9fe643eaaa2278beb3b7f7daab55bd649e1449552"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.041002 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" event={"ID":"b32201a1-e841-4c4c-8988-0b866243b255","Type":"ContainerStarted","Data":"cef7348f131b98393261b52cb7d2dee05df4b90ff13cfaadeab8a7702125b586"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.043405 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" event={"ID":"b2ca2442-7e23-40a8-9ca1-1ef11251d34d","Type":"ContainerStarted","Data":"9ff688f9cc24cbd6ae500cec21391a23e06139476297b30405978ec1c872a9a3"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.046194 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" event={"ID":"25cc3669-bdfd-44eb-83d0-85ca303250ea","Type":"ContainerStarted","Data":"a0cd3c3ac73a5944a6face648a2e2c95a7b42983f0cfd8d3bf408651d9ca0397"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.047512 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" event={"ID":"5c314cb3-27af-4a08-b96b-91f9a70f7524","Type":"ContainerStarted","Data":"101d27b1e99f5d615cac994a2d2eec9f534659a148b41982e6be0c9f6fee06af"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.048874 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" event={"ID":"43c420bf-cf21-4caa-90ce-0a9009d3bc45","Type":"ContainerStarted","Data":"0d3c83ff7ddc67235bef9f18692db876a5b6e82112e90f12a97fb7d33a87f247"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.053027 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" podStartSLOduration=4.132145762 podStartE2EDuration="17.053014304s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.366475314 +0000 UTC m=+1051.720390215" lastFinishedPulling="2025-10-08 15:30:55.287343866 +0000 UTC m=+1064.641258757" observedRunningTime="2025-10-08 15:30:57.052206164 +0000 UTC m=+1066.406121065" watchObservedRunningTime="2025-10-08 15:30:57.053014304 +0000 UTC m=+1066.406929205" Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.056835 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" event={"ID":"f8b50124-bf7a-497b-bdc6-42bbbe564a7d","Type":"ContainerStarted","Data":"3e6e434902abe68fe692d27dc8e0ce59eb93f07c757aedec311aec0a0ab11117"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.057098 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.058787 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" event={"ID":"2747410b-55fa-4a7f-9146-5662ee48cce7","Type":"ContainerStarted","Data":"888ac9d5abe79acdc6f748658fdbdd31f88c5c3a0fa446c7884d6deab70e9d3b"} Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.084162 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" podStartSLOduration=3.63454121 podStartE2EDuration="17.084146791s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:41.748861041 +0000 UTC m=+1051.102775942" lastFinishedPulling="2025-10-08 15:30:55.198466622 +0000 UTC m=+1064.552381523" observedRunningTime="2025-10-08 15:30:57.07937511 +0000 UTC m=+1066.433290011" watchObservedRunningTime="2025-10-08 15:30:57.084146791 +0000 UTC m=+1066.438061692" Oct 08 15:30:57 crc kubenswrapper[4945]: I1008 15:30:57.112991 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" podStartSLOduration=4.244186992 podStartE2EDuration="17.112976059s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.404518695 +0000 UTC m=+1051.758433596" lastFinishedPulling="2025-10-08 15:30:55.273307762 +0000 UTC m=+1064.627222663" observedRunningTime="2025-10-08 15:30:57.107146292 +0000 UTC m=+1066.461061193" watchObservedRunningTime="2025-10-08 15:30:57.112976059 +0000 UTC m=+1066.466890960" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.071518 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" event={"ID":"2747410b-55fa-4a7f-9146-5662ee48cce7","Type":"ContainerStarted","Data":"3926d1698df866afb8e9b4ff641eb72c30fbe01d4df934c119fd27408f94abe6"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.072354 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.076604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" event={"ID":"5c314cb3-27af-4a08-b96b-91f9a70f7524","Type":"ContainerStarted","Data":"ca0024aa81a56a30424a98eac59af27427ac33c1e7409732f39ebf2c3a6f5d44"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.077054 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.080624 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" event={"ID":"25cc3669-bdfd-44eb-83d0-85ca303250ea","Type":"ContainerStarted","Data":"00c26114ca66465a53b6d60a0d8caaee6fdbdbca98c58a221881c6cd3efddd69"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.081010 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.082842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" event={"ID":"c3ae7aef-2f32-4d2d-b8be-e29dd91ede71","Type":"ContainerStarted","Data":"02be4ad5770f3ff5921206fe9f74ea07161b5d4de0059fa1ce42a153671bb6dc"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.083212 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.084852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" event={"ID":"43c420bf-cf21-4caa-90ce-0a9009d3bc45","Type":"ContainerStarted","Data":"def1e5449c1864ead7afe5d1e960e23aef4fd1a7d7794e8b78d9998be7309994"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.085217 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.088055 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" event={"ID":"b2ca2442-7e23-40a8-9ca1-1ef11251d34d","Type":"ContainerStarted","Data":"192772c4fa1b81f70097d2aa631ecea31728168f5baa1fca64cc5238827ff604"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.088474 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.091941 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" podStartSLOduration=5.551116346 podStartE2EDuration="19.091926771s" podCreationTimestamp="2025-10-08 15:30:39 +0000 UTC" firstStartedPulling="2025-10-08 15:30:41.759176101 +0000 UTC m=+1051.113091002" lastFinishedPulling="2025-10-08 15:30:55.299986516 +0000 UTC m=+1064.653901427" observedRunningTime="2025-10-08 15:30:58.088615218 +0000 UTC m=+1067.442530119" watchObservedRunningTime="2025-10-08 15:30:58.091926771 +0000 UTC m=+1067.445841672" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.098494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" event={"ID":"9555aa6e-7b8f-4354-96e8-e639719e1b56","Type":"ContainerStarted","Data":"d4bde1ad9d0433e111756cad07a9ca9e212d81645e00db8ce7ee78082c086d56"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.099079 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.107580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" event={"ID":"400ef757-79e4-4dd6-8c0a-2440835ac048","Type":"ContainerStarted","Data":"487e2b646d21cdccb287030b5fe843d54c2dff9c5742d37017aa585756bf4456"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.108269 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.112244 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" podStartSLOduration=4.984258779 podStartE2EDuration="18.112222645s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.172506093 +0000 UTC m=+1051.526420994" lastFinishedPulling="2025-10-08 15:30:55.300469959 +0000 UTC m=+1064.654384860" observedRunningTime="2025-10-08 15:30:58.103041182 +0000 UTC m=+1067.456956073" watchObservedRunningTime="2025-10-08 15:30:58.112222645 +0000 UTC m=+1067.466137546" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.127344 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" podStartSLOduration=5.221481954 podStartE2EDuration="18.127323236s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.395420336 +0000 UTC m=+1051.749335227" lastFinishedPulling="2025-10-08 15:30:55.301261598 +0000 UTC m=+1064.655176509" observedRunningTime="2025-10-08 15:30:58.119019886 +0000 UTC m=+1067.472934787" watchObservedRunningTime="2025-10-08 15:30:58.127323236 +0000 UTC m=+1067.481238137" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.127887 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" event={"ID":"344902f1-cf0c-4390-8f2c-2fe41cf9ccd9","Type":"ContainerStarted","Data":"4f341c2fc35849dac1a05400d0c93a9948aa22bccb87c3a30724b8a84785842b"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.129604 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.139273 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" event={"ID":"441d050d-f359-46d8-8c29-0b1506c1f240","Type":"ContainerStarted","Data":"de8206f8c510ecca335997124c609bd827658fe5d4e2361180de57b1d3e48bc4"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.140055 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.142423 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" podStartSLOduration=4.702340027 podStartE2EDuration="18.142402676s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:41.758911925 +0000 UTC m=+1051.112826826" lastFinishedPulling="2025-10-08 15:30:55.198974574 +0000 UTC m=+1064.552889475" observedRunningTime="2025-10-08 15:30:58.136907988 +0000 UTC m=+1067.490822909" watchObservedRunningTime="2025-10-08 15:30:58.142402676 +0000 UTC m=+1067.496317577" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.145600 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" event={"ID":"a8d00f8c-42c6-4d99-bb76-bdd491597e3a","Type":"ContainerStarted","Data":"a8411d930fe10594cce7876c57e5a4bddba086aa1ce384236ac8762e623429d4"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.146428 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.156540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" event={"ID":"b32201a1-e841-4c4c-8988-0b866243b255","Type":"ContainerStarted","Data":"af0d0d8c626dda21b3201e93e15a3ed4ece5cf27702c3f104c2d132fe3f10d73"} Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.156953 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.159879 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" podStartSLOduration=5.166005492 podStartE2EDuration="18.159860248s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.30613062 +0000 UTC m=+1051.660045511" lastFinishedPulling="2025-10-08 15:30:55.299985366 +0000 UTC m=+1064.653900267" observedRunningTime="2025-10-08 15:30:58.151963818 +0000 UTC m=+1067.505878719" watchObservedRunningTime="2025-10-08 15:30:58.159860248 +0000 UTC m=+1067.513775149" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.197015 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" podStartSLOduration=5.189158926 podStartE2EDuration="18.196998106s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.302735174 +0000 UTC m=+1051.656650075" lastFinishedPulling="2025-10-08 15:30:55.310574354 +0000 UTC m=+1064.664489255" observedRunningTime="2025-10-08 15:30:58.191970739 +0000 UTC m=+1067.545885630" watchObservedRunningTime="2025-10-08 15:30:58.196998106 +0000 UTC m=+1067.550913007" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.199184 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" podStartSLOduration=5.2849124960000005 podStartE2EDuration="18.199175931s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.404238188 +0000 UTC m=+1051.758153089" lastFinishedPulling="2025-10-08 15:30:55.318501623 +0000 UTC m=+1064.672416524" observedRunningTime="2025-10-08 15:30:58.173150493 +0000 UTC m=+1067.527065394" watchObservedRunningTime="2025-10-08 15:30:58.199175931 +0000 UTC m=+1067.553090832" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.215428 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" podStartSLOduration=4.483393037 podStartE2EDuration="18.215412702s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:41.570018563 +0000 UTC m=+1050.923933464" lastFinishedPulling="2025-10-08 15:30:55.302038208 +0000 UTC m=+1064.655953129" observedRunningTime="2025-10-08 15:30:58.209717427 +0000 UTC m=+1067.563632338" watchObservedRunningTime="2025-10-08 15:30:58.215412702 +0000 UTC m=+1067.569327603" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.233556 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" podStartSLOduration=5.106631631 podStartE2EDuration="18.233541979s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.186675542 +0000 UTC m=+1051.540590443" lastFinishedPulling="2025-10-08 15:30:55.31358589 +0000 UTC m=+1064.667500791" observedRunningTime="2025-10-08 15:30:58.232963885 +0000 UTC m=+1067.586878806" watchObservedRunningTime="2025-10-08 15:30:58.233541979 +0000 UTC m=+1067.587456880" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.265783 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" podStartSLOduration=6.03205182 podStartE2EDuration="18.265763263s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:43.068565991 +0000 UTC m=+1052.422480892" lastFinishedPulling="2025-10-08 15:30:55.302277434 +0000 UTC m=+1064.656192335" observedRunningTime="2025-10-08 15:30:58.264868061 +0000 UTC m=+1067.618782972" watchObservedRunningTime="2025-10-08 15:30:58.265763263 +0000 UTC m=+1067.619678164" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.292974 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" podStartSLOduration=5.348901522 podStartE2EDuration="18.29293399s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.343316829 +0000 UTC m=+1051.697231730" lastFinishedPulling="2025-10-08 15:30:55.287349297 +0000 UTC m=+1064.641264198" observedRunningTime="2025-10-08 15:30:58.283503971 +0000 UTC m=+1067.637418872" watchObservedRunningTime="2025-10-08 15:30:58.29293399 +0000 UTC m=+1067.646848891" Oct 08 15:30:58 crc kubenswrapper[4945]: I1008 15:30:58.309984 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" podStartSLOduration=5.302256494 podStartE2EDuration="18.30995632s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.29388319 +0000 UTC m=+1051.647798081" lastFinishedPulling="2025-10-08 15:30:55.301583006 +0000 UTC m=+1064.655497907" observedRunningTime="2025-10-08 15:30:58.300617544 +0000 UTC m=+1067.654532445" watchObservedRunningTime="2025-10-08 15:30:58.30995632 +0000 UTC m=+1067.663871221" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.176816 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-5t8fx" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.177182 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-lr6hn" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.177210 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qxbqr" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.182386 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-4fjrf" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.182539 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-tb4rv" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.185690 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-7b8pm" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.189592 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-h27l8" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.458363 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-62hpj" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.754843 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-lwppj" Oct 08 15:31:00 crc kubenswrapper[4945]: I1008 15:31:00.926618 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-rn55t" Oct 08 15:31:02 crc kubenswrapper[4945]: I1008 15:31:02.330150 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.209061 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" event={"ID":"efd39ce1-ac6e-449d-b32f-fc451ffde97e","Type":"ContainerStarted","Data":"7f1799ec82a05c1ec18839327b66f8170c893f74270bab2bb8f546c38539cee6"} Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.209638 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.211021 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" event={"ID":"cbe3a47b-033a-410d-b2be-bd38e4b9c0ab","Type":"ContainerStarted","Data":"1e477b0a3a7648d07392edb8ba4f0d7eadc4eaec1b69ee3adc19f2b20e122cf7"} Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.211151 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.212449 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" event={"ID":"28b9d7cf-ad46-4c2a-a942-f7b20c054216","Type":"ContainerStarted","Data":"fd48f0a781c6610b33c375652ea8c22eb72080612ce5e86ad48ebc55c0656ae7"} Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.212603 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.214038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" event={"ID":"f11222c0-9d0f-4d3e-945f-b991135e4235","Type":"ContainerStarted","Data":"8458d5f8444a0747e54038ca55fb8f15e6a4240f01157a81b1d1ee1aa0f85de6"} Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.214270 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.216459 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" event={"ID":"1d068bb6-a523-4903-95b7-7c62862c2bb3","Type":"ContainerStarted","Data":"3e083637b104ebcb8cb0d494f2903efd0f9933ad094b024247394f7d4d4c4c32"} Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.218038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" event={"ID":"1112a081-2d10-48e2-b345-8ef841c5dc50","Type":"ContainerStarted","Data":"37f0d43abed964428a6084951b6a4d850243e117de1b0a77f01c7d3dc08ea246"} Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.218182 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.220205 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" event={"ID":"e0437b00-8fcc-412a-90fa-32ea56c71421","Type":"ContainerStarted","Data":"8c5fc108eec9f185933c13436c57bc694f1945bc231315698cc663fc9bf9c266"} Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.220400 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.233300 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" podStartSLOduration=3.842868045 podStartE2EDuration="25.23327662s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.424788078 +0000 UTC m=+1051.778702979" lastFinishedPulling="2025-10-08 15:31:03.815196643 +0000 UTC m=+1073.169111554" observedRunningTime="2025-10-08 15:31:05.226232192 +0000 UTC m=+1074.580147123" watchObservedRunningTime="2025-10-08 15:31:05.23327662 +0000 UTC m=+1074.587191521" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.261394 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" podStartSLOduration=3.965260755 podStartE2EDuration="25.261361249s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.518371561 +0000 UTC m=+1051.872286462" lastFinishedPulling="2025-10-08 15:31:03.814472045 +0000 UTC m=+1073.168386956" observedRunningTime="2025-10-08 15:31:05.241386615 +0000 UTC m=+1074.595301516" watchObservedRunningTime="2025-10-08 15:31:05.261361249 +0000 UTC m=+1074.615276180" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.276315 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" podStartSLOduration=3.900781897 podStartE2EDuration="25.276297676s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.465268531 +0000 UTC m=+1051.819183432" lastFinishedPulling="2025-10-08 15:31:03.84078431 +0000 UTC m=+1073.194699211" observedRunningTime="2025-10-08 15:31:05.271333671 +0000 UTC m=+1074.625248572" watchObservedRunningTime="2025-10-08 15:31:05.276297676 +0000 UTC m=+1074.630212587" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.291873 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx" podStartSLOduration=3.898195923 podStartE2EDuration="25.29185181s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.518773352 +0000 UTC m=+1051.872688253" lastFinishedPulling="2025-10-08 15:31:03.912429229 +0000 UTC m=+1073.266344140" observedRunningTime="2025-10-08 15:31:05.287019567 +0000 UTC m=+1074.640934488" watchObservedRunningTime="2025-10-08 15:31:05.29185181 +0000 UTC m=+1074.645766711" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.306114 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" podStartSLOduration=3.906189754 podStartE2EDuration="25.306090249s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.424954712 +0000 UTC m=+1051.778869613" lastFinishedPulling="2025-10-08 15:31:03.824855207 +0000 UTC m=+1073.178770108" observedRunningTime="2025-10-08 15:31:05.303791301 +0000 UTC m=+1074.657706202" watchObservedRunningTime="2025-10-08 15:31:05.306090249 +0000 UTC m=+1074.660005150" Oct 08 15:31:05 crc kubenswrapper[4945]: I1008 15:31:05.323054 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" podStartSLOduration=3.993350925 podStartE2EDuration="25.323036587s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.484743102 +0000 UTC m=+1051.838658003" lastFinishedPulling="2025-10-08 15:31:03.814428764 +0000 UTC m=+1073.168343665" observedRunningTime="2025-10-08 15:31:05.31878364 +0000 UTC m=+1074.672698561" watchObservedRunningTime="2025-10-08 15:31:05.323036587 +0000 UTC m=+1074.676951488" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.578904 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-hk5rr" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.597324 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" podStartSLOduration=9.241675359 podStartE2EDuration="30.597301716s" podCreationTimestamp="2025-10-08 15:30:40 +0000 UTC" firstStartedPulling="2025-10-08 15:30:42.485015169 +0000 UTC m=+1051.838930070" lastFinishedPulling="2025-10-08 15:31:03.840641516 +0000 UTC m=+1073.194556427" observedRunningTime="2025-10-08 15:31:05.336481836 +0000 UTC m=+1074.690396737" watchObservedRunningTime="2025-10-08 15:31:10.597301716 +0000 UTC m=+1079.951216637" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.615457 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ht7cm" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.625832 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-p7bfn" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.721507 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-kx5k7" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.791447 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j9vft" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.839139 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-c8h57" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.864262 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-r9mp8" Oct 08 15:31:10 crc kubenswrapper[4945]: I1008 15:31:10.960696 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-t56f2" Oct 08 15:31:11 crc kubenswrapper[4945]: I1008 15:31:11.008733 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-df6684d58-22z48" Oct 08 15:31:11 crc kubenswrapper[4945]: I1008 15:31:11.142804 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-dwfj9" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.407967 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c59bb4f7-2s665"] Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.409601 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.413306 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4h8wk" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.414762 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.414987 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.415427 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.420809 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c59bb4f7-2s665"] Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.463482 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b6644b45c-8cvgw"] Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.464885 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.467761 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.474893 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6644b45c-8cvgw"] Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.495015 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrg5l\" (UniqueName: \"kubernetes.io/projected/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-kube-api-access-mrg5l\") pod \"dnsmasq-dns-7c59bb4f7-2s665\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.495072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-config\") pod \"dnsmasq-dns-7c59bb4f7-2s665\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.596547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-config\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.596603 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-dns-svc\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.596631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrg5l\" (UniqueName: \"kubernetes.io/projected/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-kube-api-access-mrg5l\") pod \"dnsmasq-dns-7c59bb4f7-2s665\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.596648 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql4fd\" (UniqueName: \"kubernetes.io/projected/cdede57e-b9a1-4595-bad3-f6c1355e2743-kube-api-access-ql4fd\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.596676 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-config\") pod \"dnsmasq-dns-7c59bb4f7-2s665\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.597721 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-config\") pod \"dnsmasq-dns-7c59bb4f7-2s665\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.614169 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrg5l\" (UniqueName: \"kubernetes.io/projected/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-kube-api-access-mrg5l\") pod \"dnsmasq-dns-7c59bb4f7-2s665\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.697942 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-config\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.697998 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-dns-svc\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.698027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql4fd\" (UniqueName: \"kubernetes.io/projected/cdede57e-b9a1-4595-bad3-f6c1355e2743-kube-api-access-ql4fd\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.698863 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-dns-svc\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.698914 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-config\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.712214 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql4fd\" (UniqueName: \"kubernetes.io/projected/cdede57e-b9a1-4595-bad3-f6c1355e2743-kube-api-access-ql4fd\") pod \"dnsmasq-dns-b6644b45c-8cvgw\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.735120 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:31 crc kubenswrapper[4945]: I1008 15:31:31.782862 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:32 crc kubenswrapper[4945]: I1008 15:31:32.198556 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c59bb4f7-2s665"] Oct 08 15:31:32 crc kubenswrapper[4945]: I1008 15:31:32.277199 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6644b45c-8cvgw"] Oct 08 15:31:32 crc kubenswrapper[4945]: W1008 15:31:32.280060 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdede57e_b9a1_4595_bad3_f6c1355e2743.slice/crio-2a07c75496da1b63083db51ffa2997cde4c23bbbd6f17602cf8f32575f23948a WatchSource:0}: Error finding container 2a07c75496da1b63083db51ffa2997cde4c23bbbd6f17602cf8f32575f23948a: Status 404 returned error can't find the container with id 2a07c75496da1b63083db51ffa2997cde4c23bbbd6f17602cf8f32575f23948a Oct 08 15:31:32 crc kubenswrapper[4945]: I1008 15:31:32.447850 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" event={"ID":"cdede57e-b9a1-4595-bad3-f6c1355e2743","Type":"ContainerStarted","Data":"2a07c75496da1b63083db51ffa2997cde4c23bbbd6f17602cf8f32575f23948a"} Oct 08 15:31:32 crc kubenswrapper[4945]: I1008 15:31:32.449121 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" event={"ID":"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5","Type":"ContainerStarted","Data":"2e5811513cf2d2cf3e2f6be12eb04dc27795057a88461d531e4dddf60dcd188c"} Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.305496 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c59bb4f7-2s665"] Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.331687 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-669c8b4d45-5p7x4"] Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.333046 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.344794 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-669c8b4d45-5p7x4"] Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.460038 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-dns-svc\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.460129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl4st\" (UniqueName: \"kubernetes.io/projected/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-kube-api-access-tl4st\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.460260 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-config\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.561684 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl4st\" (UniqueName: \"kubernetes.io/projected/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-kube-api-access-tl4st\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.561819 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-config\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.561869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-dns-svc\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.562932 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-config\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.562967 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-dns-svc\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.581287 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl4st\" (UniqueName: \"kubernetes.io/projected/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-kube-api-access-tl4st\") pod \"dnsmasq-dns-669c8b4d45-5p7x4\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.627608 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6644b45c-8cvgw"] Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.649825 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.658993 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78bdd78b69-hpwh4"] Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.662156 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.668061 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78bdd78b69-hpwh4"] Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.764568 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-config\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.764616 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9gn8\" (UniqueName: \"kubernetes.io/projected/098407b4-11e2-40a3-b3d1-b47b9d9ca618-kube-api-access-c9gn8\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.764790 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-dns-svc\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.868082 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-config\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.868169 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9gn8\" (UniqueName: \"kubernetes.io/projected/098407b4-11e2-40a3-b3d1-b47b9d9ca618-kube-api-access-c9gn8\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.868227 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-dns-svc\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.869154 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-config\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.869267 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-dns-svc\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.889167 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9gn8\" (UniqueName: \"kubernetes.io/projected/098407b4-11e2-40a3-b3d1-b47b9d9ca618-kube-api-access-c9gn8\") pod \"dnsmasq-dns-78bdd78b69-hpwh4\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.945897 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-669c8b4d45-5p7x4"] Oct 08 15:31:35 crc kubenswrapper[4945]: I1008 15:31:35.981431 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.007540 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9bb9d48f5-99b2z"] Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.012847 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.051014 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9bb9d48f5-99b2z"] Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.175364 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npwpf\" (UniqueName: \"kubernetes.io/projected/8316db8b-1cdf-4494-a07d-2552ff4e72e4-kube-api-access-npwpf\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.175423 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-config\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.175521 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-dns-svc\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.276638 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-dns-svc\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.276728 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npwpf\" (UniqueName: \"kubernetes.io/projected/8316db8b-1cdf-4494-a07d-2552ff4e72e4-kube-api-access-npwpf\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.276768 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-config\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.277587 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-config\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.277750 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-dns-svc\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.292805 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npwpf\" (UniqueName: \"kubernetes.io/projected/8316db8b-1cdf-4494-a07d-2552ff4e72e4-kube-api-access-npwpf\") pod \"dnsmasq-dns-9bb9d48f5-99b2z\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.350172 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.496596 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.500776 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.502349 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.503485 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.503909 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.503914 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.505102 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.505511 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lcgsm" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.505682 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.520306 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.579583 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.579644 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.579674 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.579693 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec56f6d2-9c69-47f1-b834-169de890c512-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.579714 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.580409 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.580470 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j96z6\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-kube-api-access-j96z6\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.580502 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.580552 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec56f6d2-9c69-47f1-b834-169de890c512-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.580571 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.580939 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.682983 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683042 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j96z6\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-kube-api-access-j96z6\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683091 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec56f6d2-9c69-47f1-b834-169de890c512-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683133 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683155 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683252 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec56f6d2-9c69-47f1-b834-169de890c512-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.683755 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.684522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.684822 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.684909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.685219 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.686262 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.686440 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.691259 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.693040 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.693628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec56f6d2-9c69-47f1-b834-169de890c512-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.701231 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j96z6\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-kube-api-access-j96z6\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.705074 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec56f6d2-9c69-47f1-b834-169de890c512-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.708964 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.767612 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.768932 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.772084 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.772088 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.772165 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-sdjcl" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.772171 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.772498 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.772849 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.776404 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.781471 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.822436 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.889278 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.889373 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420a6f35-1c79-416f-a7f4-8b9f46e337bc-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.889434 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.889570 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.889685 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420a6f35-1c79-416f-a7f4-8b9f46e337bc-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.889891 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.889951 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.890047 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc4s7\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-kube-api-access-fc4s7\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.890170 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.890268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.890317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991738 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991800 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991832 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991882 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420a6f35-1c79-416f-a7f4-8b9f46e337bc-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991959 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.991995 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420a6f35-1c79-416f-a7f4-8b9f46e337bc-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.992051 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.992080 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.992131 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc4s7\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-kube-api-access-fc4s7\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.992214 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.992844 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.993313 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.995469 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.995978 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.996137 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420a6f35-1c79-416f-a7f4-8b9f46e337bc-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.996976 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:36 crc kubenswrapper[4945]: I1008 15:31:36.998816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420a6f35-1c79-416f-a7f4-8b9f46e337bc-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.001302 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.009021 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420a6f35-1c79-416f-a7f4-8b9f46e337bc-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.009544 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc4s7\" (UniqueName: \"kubernetes.io/projected/420a6f35-1c79-416f-a7f4-8b9f46e337bc-kube-api-access-fc4s7\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.012923 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"420a6f35-1c79-416f-a7f4-8b9f46e337bc\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.089328 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.189906 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.191307 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.201671 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.201854 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.202128 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.202692 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mqvcn" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.202825 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.202937 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.203079 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.203211 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctr6t\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-kube-api-access-ctr6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296079 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296130 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296168 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296186 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296207 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296224 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296242 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296275 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296290 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.296310 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.397851 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.397923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.397944 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.397971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.397993 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.398028 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.398045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.398070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.398810 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.398892 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.399352 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.399470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.399577 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.400460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctr6t\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-kube-api-access-ctr6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.400511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.400589 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.400864 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.402675 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.403950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.404136 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.404520 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.424650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctr6t\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-kube-api-access-ctr6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.429149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:37 crc kubenswrapper[4945]: I1008 15:31:37.521682 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.564448 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.566221 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.570444 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.570900 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.571419 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.571946 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.572198 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6lpp2" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.577396 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.578037 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.614031 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.615374 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.618529 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.618571 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hqvjr" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.618585 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.618640 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.619306 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.661317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.661589 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-kolla-config\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.661679 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-config-data-default\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.661765 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-secrets\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.661872 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6dvs\" (UniqueName: \"kubernetes.io/projected/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-kube-api-access-l6dvs\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.662045 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.662144 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.662199 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.662409 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763529 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763582 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763601 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763626 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763645 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-kolla-config\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763663 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-config-data-default\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-secrets\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763741 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6dvs\" (UniqueName: \"kubernetes.io/projected/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-kube-api-access-l6dvs\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763758 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763780 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763800 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763820 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763839 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763859 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763885 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2w9r\" (UniqueName: \"kubernetes.io/projected/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-kube-api-access-t2w9r\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.763946 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.764492 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.764797 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.765054 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-config-data-default\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.765233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.765460 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-kolla-config\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.772832 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.774474 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-secrets\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.774857 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.779430 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6dvs\" (UniqueName: \"kubernetes.io/projected/88f0eb1e-a2bc-4b2f-b14d-8cef0261334e-kube-api-access-l6dvs\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.786898 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e\") " pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867174 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867289 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867318 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867347 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2w9r\" (UniqueName: \"kubernetes.io/projected/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-kube-api-access-t2w9r\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867369 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867390 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867439 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867499 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.867515 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.868042 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.868422 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.869149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.869307 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.869821 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.871751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.875431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.875859 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.897773 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2w9r\" (UniqueName: \"kubernetes.io/projected/1d5103f7-4fab-43a6-8989-1b505ac5d8d0-kube-api-access-t2w9r\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.900320 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.901199 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1d5103f7-4fab-43a6-8989-1b505ac5d8d0\") " pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:40 crc kubenswrapper[4945]: I1008 15:31:40.933973 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.069900 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.071446 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.073475 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.074341 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.078419 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-phrjs" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.089701 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.171232 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.171573 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp7rv\" (UniqueName: \"kubernetes.io/projected/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-kube-api-access-tp7rv\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.171601 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.171649 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-config-data\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.171935 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-kolla-config\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.273730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.273781 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp7rv\" (UniqueName: \"kubernetes.io/projected/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-kube-api-access-tp7rv\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.273798 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.273814 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-config-data\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.273878 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-kolla-config\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.274771 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-kolla-config\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.274848 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-config-data\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.277826 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.278559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.296639 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp7rv\" (UniqueName: \"kubernetes.io/projected/ceb183ce-9008-4a1a-8b6b-26d47b2b246d-kube-api-access-tp7rv\") pod \"memcached-0\" (UID: \"ceb183ce-9008-4a1a-8b6b-26d47b2b246d\") " pod="openstack/memcached-0" Oct 08 15:31:41 crc kubenswrapper[4945]: I1008 15:31:41.386150 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 15:31:42 crc kubenswrapper[4945]: I1008 15:31:42.857828 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:31:42 crc kubenswrapper[4945]: I1008 15:31:42.894595 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:31:42 crc kubenswrapper[4945]: I1008 15:31:42.894703 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 15:31:42 crc kubenswrapper[4945]: I1008 15:31:42.898133 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-gtl46" Oct 08 15:31:43 crc kubenswrapper[4945]: I1008 15:31:43.016917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p6s9\" (UniqueName: \"kubernetes.io/projected/4971563b-eba1-4306-a525-e088a5b3f754-kube-api-access-8p6s9\") pod \"kube-state-metrics-0\" (UID: \"4971563b-eba1-4306-a525-e088a5b3f754\") " pod="openstack/kube-state-metrics-0" Oct 08 15:31:43 crc kubenswrapper[4945]: I1008 15:31:43.118514 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p6s9\" (UniqueName: \"kubernetes.io/projected/4971563b-eba1-4306-a525-e088a5b3f754-kube-api-access-8p6s9\") pod \"kube-state-metrics-0\" (UID: \"4971563b-eba1-4306-a525-e088a5b3f754\") " pod="openstack/kube-state-metrics-0" Oct 08 15:31:43 crc kubenswrapper[4945]: I1008 15:31:43.147473 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p6s9\" (UniqueName: \"kubernetes.io/projected/4971563b-eba1-4306-a525-e088a5b3f754-kube-api-access-8p6s9\") pod \"kube-state-metrics-0\" (UID: \"4971563b-eba1-4306-a525-e088a5b3f754\") " pod="openstack/kube-state-metrics-0" Oct 08 15:31:43 crc kubenswrapper[4945]: I1008 15:31:43.231761 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.185380 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.187259 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.190374 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.190547 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.190665 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.191488 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-qmfdz" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.191944 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.198134 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.198137 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.237794 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.237882 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.237917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q9k6\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-kube-api-access-8q9k6\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.237955 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.237983 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.238003 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.238022 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.238044 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.339631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q9k6\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-kube-api-access-8q9k6\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.339737 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.339966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.339998 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.340036 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.340075 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.340138 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.340233 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.341583 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.345762 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.346016 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9dcd0c273e01834b6d6b9a81fac4d1c0f2ca078125a1dee333b69bb46fb7d5e3/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.346085 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.346171 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.347513 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.349632 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.351060 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.370946 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q9k6\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-kube-api-access-8q9k6\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.393320 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:44 crc kubenswrapper[4945]: I1008 15:31:44.517977 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.533001 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cwwpc"] Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.534679 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.537599 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-qx8l4" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.545766 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.546066 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.551736 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-67zmn"] Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.558055 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.570203 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cwwpc"] Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.578941 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-run-ovn\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.579005 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b70322d-ae60-46b4-828b-1f4a01cae6b6-scripts\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.579051 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b70322d-ae60-46b4-828b-1f4a01cae6b6-ovn-controller-tls-certs\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.579133 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcgkd\" (UniqueName: \"kubernetes.io/projected/1b70322d-ae60-46b4-828b-1f4a01cae6b6-kube-api-access-fcgkd\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.579225 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-run\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.579271 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-log-ovn\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.579352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b70322d-ae60-46b4-828b-1f4a01cae6b6-combined-ca-bundle\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.593706 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-67zmn"] Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.680900 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-run-ovn\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.680985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b70322d-ae60-46b4-828b-1f4a01cae6b6-scripts\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681024 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b70322d-ae60-46b4-828b-1f4a01cae6b6-ovn-controller-tls-certs\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681074 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-etc-ovs\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681095 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fa35d9f-678d-4820-9be9-7db07c5ffd07-scripts\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcgkd\" (UniqueName: \"kubernetes.io/projected/1b70322d-ae60-46b4-828b-1f4a01cae6b6-kube-api-access-fcgkd\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-run\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681172 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-log-ovn\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681199 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-log\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681219 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-lib\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681252 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm9lm\" (UniqueName: \"kubernetes.io/projected/7fa35d9f-678d-4820-9be9-7db07c5ffd07-kube-api-access-vm9lm\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681275 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b70322d-ae60-46b4-828b-1f4a01cae6b6-combined-ca-bundle\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681312 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-run\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681477 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-run-ovn\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681602 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-run\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.681720 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1b70322d-ae60-46b4-828b-1f4a01cae6b6-var-log-ovn\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.683001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b70322d-ae60-46b4-828b-1f4a01cae6b6-scripts\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.685334 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b70322d-ae60-46b4-828b-1f4a01cae6b6-ovn-controller-tls-certs\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.687653 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b70322d-ae60-46b4-828b-1f4a01cae6b6-combined-ca-bundle\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.697989 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcgkd\" (UniqueName: \"kubernetes.io/projected/1b70322d-ae60-46b4-828b-1f4a01cae6b6-kube-api-access-fcgkd\") pod \"ovn-controller-cwwpc\" (UID: \"1b70322d-ae60-46b4-828b-1f4a01cae6b6\") " pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782294 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-etc-ovs\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782344 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fa35d9f-678d-4820-9be9-7db07c5ffd07-scripts\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782621 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-log\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782689 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-lib\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782748 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-etc-ovs\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782846 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm9lm\" (UniqueName: \"kubernetes.io/projected/7fa35d9f-678d-4820-9be9-7db07c5ffd07-kube-api-access-vm9lm\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-run\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782987 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-lib\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.782879 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-log\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.783306 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fa35d9f-678d-4820-9be9-7db07c5ffd07-var-run\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.784366 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fa35d9f-678d-4820-9be9-7db07c5ffd07-scripts\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.800434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm9lm\" (UniqueName: \"kubernetes.io/projected/7fa35d9f-678d-4820-9be9-7db07c5ffd07-kube-api-access-vm9lm\") pod \"ovn-controller-ovs-67zmn\" (UID: \"7fa35d9f-678d-4820-9be9-7db07c5ffd07\") " pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.860682 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cwwpc" Oct 08 15:31:46 crc kubenswrapper[4945]: I1008 15:31:46.885210 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.430318 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.439659 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.439758 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.441932 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.442079 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-28vfh" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.443167 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.443618 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.443786 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493287 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493356 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493617 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-config\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493692 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493789 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493840 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493874 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.493912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffjvn\" (UniqueName: \"kubernetes.io/projected/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-kube-api-access-ffjvn\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595741 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595792 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595844 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-config\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595872 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595906 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595928 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595950 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.595966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffjvn\" (UniqueName: \"kubernetes.io/projected/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-kube-api-access-ffjvn\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.596896 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.596929 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-config\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.597347 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.597346 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.601213 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.601621 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.602062 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.618862 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffjvn\" (UniqueName: \"kubernetes.io/projected/c4cb5094-ffbc-4fff-bbc4-0a3a35067782-kube-api-access-ffjvn\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.625215 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c4cb5094-ffbc-4fff-bbc4-0a3a35067782\") " pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:47 crc kubenswrapper[4945]: I1008 15:31:47.772702 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 15:31:50 crc kubenswrapper[4945]: I1008 15:31:50.973239 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 15:31:50 crc kubenswrapper[4945]: I1008 15:31:50.975233 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:50 crc kubenswrapper[4945]: I1008 15:31:50.982469 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 08 15:31:50 crc kubenswrapper[4945]: I1008 15:31:50.984105 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 08 15:31:50 crc kubenswrapper[4945]: I1008 15:31:50.988068 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 08 15:31:50 crc kubenswrapper[4945]: I1008 15:31:50.988428 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jkwfc" Oct 08 15:31:50 crc kubenswrapper[4945]: I1008 15:31:50.994118 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052539 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052585 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec9d8fdb-5e82-4cb6-811b-053a28588453-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052626 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec9d8fdb-5e82-4cb6-811b-053a28588453-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052645 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9d8fdb-5e82-4cb6-811b-053a28588453-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052670 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052700 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jncnq\" (UniqueName: \"kubernetes.io/projected/ec9d8fdb-5e82-4cb6-811b-053a28588453-kube-api-access-jncnq\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.052925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.154799 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.155246 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec9d8fdb-5e82-4cb6-811b-053a28588453-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.155184 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.155707 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec9d8fdb-5e82-4cb6-811b-053a28588453-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.156840 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec9d8fdb-5e82-4cb6-811b-053a28588453-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.155782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec9d8fdb-5e82-4cb6-811b-053a28588453-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.158394 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9d8fdb-5e82-4cb6-811b-053a28588453-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.158461 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.158533 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.158563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jncnq\" (UniqueName: \"kubernetes.io/projected/ec9d8fdb-5e82-4cb6-811b-053a28588453-kube-api-access-jncnq\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.158671 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.158949 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9d8fdb-5e82-4cb6-811b-053a28588453-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.165038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.165039 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.180408 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.180816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9d8fdb-5e82-4cb6-811b-053a28588453-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.188312 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jncnq\" (UniqueName: \"kubernetes.io/projected/ec9d8fdb-5e82-4cb6-811b-053a28588453-kube-api-access-jncnq\") pod \"ovsdbserver-sb-0\" (UID: \"ec9d8fdb-5e82-4cb6-811b-053a28588453\") " pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:51 crc kubenswrapper[4945]: I1008 15:31:51.296479 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 15:31:55 crc kubenswrapper[4945]: I1008 15:31:55.878123 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.177616 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.177682 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.177814 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.163:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ql4fd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-b6644b45c-8cvgw_openstack(cdede57e-b9a1-4595-bad3-f6c1355e2743): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.179005 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" podUID="cdede57e-b9a1-4595-bad3-f6c1355e2743" Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.213634 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.213688 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.213827 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.163:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mrg5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7c59bb4f7-2s665_openstack(6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:31:56 crc kubenswrapper[4945]: E1008 15:31:56.215063 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" podUID="6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5" Oct 08 15:31:56 crc kubenswrapper[4945]: I1008 15:31:56.656933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"420a6f35-1c79-416f-a7f4-8b9f46e337bc","Type":"ContainerStarted","Data":"058509d523e877c42690073a87aabd26db80ec540a4eb2b420995ee6db7b748e"} Oct 08 15:31:56 crc kubenswrapper[4945]: I1008 15:31:56.796703 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:31:56 crc kubenswrapper[4945]: I1008 15:31:56.804683 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-669c8b4d45-5p7x4"] Oct 08 15:31:56 crc kubenswrapper[4945]: W1008 15:31:56.809936 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb088ba68_7804_4b57_a1fc_5442e6bf1ef3.slice/crio-4c5f66e2f1f01513e5c89f4f14388459cebc26b0dcfd1c66d90a1740c05ad316 WatchSource:0}: Error finding container 4c5f66e2f1f01513e5c89f4f14388459cebc26b0dcfd1c66d90a1740c05ad316: Status 404 returned error can't find the container with id 4c5f66e2f1f01513e5c89f4f14388459cebc26b0dcfd1c66d90a1740c05ad316 Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.226403 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.277542 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.288271 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrg5l\" (UniqueName: \"kubernetes.io/projected/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-kube-api-access-mrg5l\") pod \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.288339 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-config\") pod \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\" (UID: \"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5\") " Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.288827 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-config" (OuterVolumeSpecName: "config") pod "6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5" (UID: "6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.293211 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-kube-api-access-mrg5l" (OuterVolumeSpecName: "kube-api-access-mrg5l") pod "6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5" (UID: "6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5"). InnerVolumeSpecName "kube-api-access-mrg5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.390198 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql4fd\" (UniqueName: \"kubernetes.io/projected/cdede57e-b9a1-4595-bad3-f6c1355e2743-kube-api-access-ql4fd\") pod \"cdede57e-b9a1-4595-bad3-f6c1355e2743\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.390263 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-config\") pod \"cdede57e-b9a1-4595-bad3-f6c1355e2743\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.390337 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-dns-svc\") pod \"cdede57e-b9a1-4595-bad3-f6c1355e2743\" (UID: \"cdede57e-b9a1-4595-bad3-f6c1355e2743\") " Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.390659 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrg5l\" (UniqueName: \"kubernetes.io/projected/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-kube-api-access-mrg5l\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.390676 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.390750 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-config" (OuterVolumeSpecName: "config") pod "cdede57e-b9a1-4595-bad3-f6c1355e2743" (UID: "cdede57e-b9a1-4595-bad3-f6c1355e2743"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.391130 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cdede57e-b9a1-4595-bad3-f6c1355e2743" (UID: "cdede57e-b9a1-4595-bad3-f6c1355e2743"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.393142 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdede57e-b9a1-4595-bad3-f6c1355e2743-kube-api-access-ql4fd" (OuterVolumeSpecName: "kube-api-access-ql4fd") pod "cdede57e-b9a1-4595-bad3-f6c1355e2743" (UID: "cdede57e-b9a1-4595-bad3-f6c1355e2743"). InnerVolumeSpecName "kube-api-access-ql4fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.457822 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.487923 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9bb9d48f5-99b2z"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.492269 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql4fd\" (UniqueName: \"kubernetes.io/projected/cdede57e-b9a1-4595-bad3-f6c1355e2743-kube-api-access-ql4fd\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.492291 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.492300 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdede57e-b9a1-4595-bad3-f6c1355e2743-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.494715 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cwwpc"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.511343 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.517351 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.522654 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.528291 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78bdd78b69-hpwh4"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.563121 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.572502 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.675397 4945 generic.go:334] "Generic (PLEG): container finished" podID="b088ba68-7804-4b57-a1fc-5442e6bf1ef3" containerID="3b6df623c6ee811b8c6585b22f83ad17d052dc2994ed87e05edaf988a8cd7b69" exitCode=0 Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.675444 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" event={"ID":"b088ba68-7804-4b57-a1fc-5442e6bf1ef3","Type":"ContainerDied","Data":"3b6df623c6ee811b8c6585b22f83ad17d052dc2994ed87e05edaf988a8cd7b69"} Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.675495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" event={"ID":"b088ba68-7804-4b57-a1fc-5442e6bf1ef3","Type":"ContainerStarted","Data":"4c5f66e2f1f01513e5c89f4f14388459cebc26b0dcfd1c66d90a1740c05ad316"} Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.679325 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.679325 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6644b45c-8cvgw" event={"ID":"cdede57e-b9a1-4595-bad3-f6c1355e2743","Type":"ContainerDied","Data":"2a07c75496da1b63083db51ffa2997cde4c23bbbd6f17602cf8f32575f23948a"} Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.680645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" event={"ID":"6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5","Type":"ContainerDied","Data":"2e5811513cf2d2cf3e2f6be12eb04dc27795057a88461d531e4dddf60dcd188c"} Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.680731 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c59bb4f7-2s665" Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.695367 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2481ce07-e2b2-4afb-b5f1-027bfefd0f38","Type":"ContainerStarted","Data":"01b5efbe7b0c8a7e069068266e7d5fc93815f4a6d6b4123ce36a261dd2ca6342"} Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.739374 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c59bb4f7-2s665"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.747236 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c59bb4f7-2s665"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.767012 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6644b45c-8cvgw"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.792132 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b6644b45c-8cvgw"] Oct 08 15:31:57 crc kubenswrapper[4945]: I1008 15:31:57.945154 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 15:31:58 crc kubenswrapper[4945]: I1008 15:31:58.035648 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5" path="/var/lib/kubelet/pods/6c06d424-bc0c-4687-a9dc-8bcfb9fc6cd5/volumes" Oct 08 15:31:58 crc kubenswrapper[4945]: I1008 15:31:58.035998 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdede57e-b9a1-4595-bad3-f6c1355e2743" path="/var/lib/kubelet/pods/cdede57e-b9a1-4595-bad3-f6c1355e2743/volumes" Oct 08 15:31:58 crc kubenswrapper[4945]: I1008 15:31:58.776664 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 15:31:58 crc kubenswrapper[4945]: I1008 15:31:58.875634 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-67zmn"] Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.923559 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b942286_9a63_4a9f_9aa9_822ad8f3a34a.slice/crio-3d1f8df3374c53cf40964f5e8a5d9cdac1321d84a9d70d13809f620a609f3544 WatchSource:0}: Error finding container 3d1f8df3374c53cf40964f5e8a5d9cdac1321d84a9d70d13809f620a609f3544: Status 404 returned error can't find the container with id 3d1f8df3374c53cf40964f5e8a5d9cdac1321d84a9d70d13809f620a609f3544 Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.937480 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8316db8b_1cdf_4494_a07d_2552ff4e72e4.slice/crio-99a710e4c9e8c848e67a9574a282222abd9f688d918e8d64abfa15c679d3e34e WatchSource:0}: Error finding container 99a710e4c9e8c848e67a9574a282222abd9f688d918e8d64abfa15c679d3e34e: Status 404 returned error can't find the container with id 99a710e4c9e8c848e67a9574a282222abd9f688d918e8d64abfa15c679d3e34e Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.941803 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d5103f7_4fab_43a6_8989_1b505ac5d8d0.slice/crio-bf7d6619c2b0fe111c00c11ba311b5c62e33f319659dee7ebd462aa53e7531dc WatchSource:0}: Error finding container bf7d6619c2b0fe111c00c11ba311b5c62e33f319659dee7ebd462aa53e7531dc: Status 404 returned error can't find the container with id bf7d6619c2b0fe111c00c11ba311b5c62e33f319659dee7ebd462aa53e7531dc Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.944327 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod098407b4_11e2_40a3_b3d1_b47b9d9ca618.slice/crio-0823821a5ccc0288c8af602dac6ec2b20bab8bdbe24ab062930c12f181cef67c WatchSource:0}: Error finding container 0823821a5ccc0288c8af602dac6ec2b20bab8bdbe24ab062930c12f181cef67c: Status 404 returned error can't find the container with id 0823821a5ccc0288c8af602dac6ec2b20bab8bdbe24ab062930c12f181cef67c Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.949779 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec56f6d2_9c69_47f1_b834_169de890c512.slice/crio-d0b11f27eb8a2cda5aa8447d50ed22c0ff12f08930346537aa8d7133cd1128a9 WatchSource:0}: Error finding container d0b11f27eb8a2cda5aa8447d50ed22c0ff12f08930346537aa8d7133cd1128a9: Status 404 returned error can't find the container with id d0b11f27eb8a2cda5aa8447d50ed22c0ff12f08930346537aa8d7133cd1128a9 Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.950355 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podceb183ce_9008_4a1a_8b6b_26d47b2b246d.slice/crio-93e51fb7901b9186b90d98a462e6460fbb54b3002e8319a83e22347d25116cac WatchSource:0}: Error finding container 93e51fb7901b9186b90d98a462e6460fbb54b3002e8319a83e22347d25116cac: Status 404 returned error can't find the container with id 93e51fb7901b9186b90d98a462e6460fbb54b3002e8319a83e22347d25116cac Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.952430 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec9d8fdb_5e82_4cb6_811b_053a28588453.slice/crio-0ee323f55054bc5f63ae4591f79d8959954353cecf484b4d2ffeccc79ba8018a WatchSource:0}: Error finding container 0ee323f55054bc5f63ae4591f79d8959954353cecf484b4d2ffeccc79ba8018a: Status 404 returned error can't find the container with id 0ee323f55054bc5f63ae4591f79d8959954353cecf484b4d2ffeccc79ba8018a Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.958571 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88f0eb1e_a2bc_4b2f_b14d_8cef0261334e.slice/crio-eebd62d06656cbf6967fcc8725557b9bd52036c12d80a40fa1dfe7ad1dbaefcb WatchSource:0}: Error finding container eebd62d06656cbf6967fcc8725557b9bd52036c12d80a40fa1dfe7ad1dbaefcb: Status 404 returned error can't find the container with id eebd62d06656cbf6967fcc8725557b9bd52036c12d80a40fa1dfe7ad1dbaefcb Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.961508 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4971563b_eba1_4306_a525_e088a5b3f754.slice/crio-6d371ce1f996cbc66dd72295bb0cf27f9de4ad398f4ed0034405d94e37acc11b WatchSource:0}: Error finding container 6d371ce1f996cbc66dd72295bb0cf27f9de4ad398f4ed0034405d94e37acc11b: Status 404 returned error can't find the container with id 6d371ce1f996cbc66dd72295bb0cf27f9de4ad398f4ed0034405d94e37acc11b Oct 08 15:31:58 crc kubenswrapper[4945]: W1008 15:31:58.963804 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4cb5094_ffbc_4fff_bbc4_0a3a35067782.slice/crio-4b0e928770aebaa19f89ea70e81acbb7bbc2f218dcbf45043aaa6f44b67c2bc0 WatchSource:0}: Error finding container 4b0e928770aebaa19f89ea70e81acbb7bbc2f218dcbf45043aaa6f44b67c2bc0: Status 404 returned error can't find the container with id 4b0e928770aebaa19f89ea70e81acbb7bbc2f218dcbf45043aaa6f44b67c2bc0 Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.144202 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.218956 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-dns-svc\") pod \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.219018 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-config\") pod \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.219098 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl4st\" (UniqueName: \"kubernetes.io/projected/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-kube-api-access-tl4st\") pod \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\" (UID: \"b088ba68-7804-4b57-a1fc-5442e6bf1ef3\") " Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.228155 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-kube-api-access-tl4st" (OuterVolumeSpecName: "kube-api-access-tl4st") pod "b088ba68-7804-4b57-a1fc-5442e6bf1ef3" (UID: "b088ba68-7804-4b57-a1fc-5442e6bf1ef3"). InnerVolumeSpecName "kube-api-access-tl4st". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.255933 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b088ba68-7804-4b57-a1fc-5442e6bf1ef3" (UID: "b088ba68-7804-4b57-a1fc-5442e6bf1ef3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.321334 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.321365 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl4st\" (UniqueName: \"kubernetes.io/projected/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-kube-api-access-tl4st\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.328196 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-config" (OuterVolumeSpecName: "config") pod "b088ba68-7804-4b57-a1fc-5442e6bf1ef3" (UID: "b088ba68-7804-4b57-a1fc-5442e6bf1ef3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.422938 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b088ba68-7804-4b57-a1fc-5442e6bf1ef3-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.712492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1d5103f7-4fab-43a6-8989-1b505ac5d8d0","Type":"ContainerStarted","Data":"bf7d6619c2b0fe111c00c11ba311b5c62e33f319659dee7ebd462aa53e7531dc"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.714425 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" event={"ID":"b088ba68-7804-4b57-a1fc-5442e6bf1ef3","Type":"ContainerDied","Data":"4c5f66e2f1f01513e5c89f4f14388459cebc26b0dcfd1c66d90a1740c05ad316"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.714460 4945 scope.go:117] "RemoveContainer" containerID="3b6df623c6ee811b8c6585b22f83ad17d052dc2994ed87e05edaf988a8cd7b69" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.714649 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669c8b4d45-5p7x4" Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.723356 4945 generic.go:334] "Generic (PLEG): container finished" podID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerID="275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235" exitCode=0 Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.723429 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" event={"ID":"098407b4-11e2-40a3-b3d1-b47b9d9ca618","Type":"ContainerDied","Data":"275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.723447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" event={"ID":"098407b4-11e2-40a3-b3d1-b47b9d9ca618","Type":"ContainerStarted","Data":"0823821a5ccc0288c8af602dac6ec2b20bab8bdbe24ab062930c12f181cef67c"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.725946 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ceb183ce-9008-4a1a-8b6b-26d47b2b246d","Type":"ContainerStarted","Data":"93e51fb7901b9186b90d98a462e6460fbb54b3002e8319a83e22347d25116cac"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.728714 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c4cb5094-ffbc-4fff-bbc4-0a3a35067782","Type":"ContainerStarted","Data":"4b0e928770aebaa19f89ea70e81acbb7bbc2f218dcbf45043aaa6f44b67c2bc0"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.730232 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerStarted","Data":"3d1f8df3374c53cf40964f5e8a5d9cdac1321d84a9d70d13809f620a609f3544"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.732046 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cwwpc" event={"ID":"1b70322d-ae60-46b4-828b-1f4a01cae6b6","Type":"ContainerStarted","Data":"abfb9616b58f3b52144528f905bee762d8e1c4a712315423be4331d2dfd7e4e5"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.733431 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec9d8fdb-5e82-4cb6-811b-053a28588453","Type":"ContainerStarted","Data":"0ee323f55054bc5f63ae4591f79d8959954353cecf484b4d2ffeccc79ba8018a"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.744036 4945 generic.go:334] "Generic (PLEG): container finished" podID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerID="d631b88dcaa75516846337ba829f3a4968b32080ee8bbb8b55b9a478e6c82488" exitCode=0 Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.744136 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" event={"ID":"8316db8b-1cdf-4494-a07d-2552ff4e72e4","Type":"ContainerDied","Data":"d631b88dcaa75516846337ba829f3a4968b32080ee8bbb8b55b9a478e6c82488"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.744157 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" event={"ID":"8316db8b-1cdf-4494-a07d-2552ff4e72e4","Type":"ContainerStarted","Data":"99a710e4c9e8c848e67a9574a282222abd9f688d918e8d64abfa15c679d3e34e"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.745248 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4971563b-eba1-4306-a525-e088a5b3f754","Type":"ContainerStarted","Data":"6d371ce1f996cbc66dd72295bb0cf27f9de4ad398f4ed0034405d94e37acc11b"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.746398 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-67zmn" event={"ID":"7fa35d9f-678d-4820-9be9-7db07c5ffd07","Type":"ContainerStarted","Data":"2068a579580ed9d2b2428e474da608486e4f4f517019e631abd67d4669890d7c"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.748861 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec56f6d2-9c69-47f1-b834-169de890c512","Type":"ContainerStarted","Data":"d0b11f27eb8a2cda5aa8447d50ed22c0ff12f08930346537aa8d7133cd1128a9"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.750445 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e","Type":"ContainerStarted","Data":"eebd62d06656cbf6967fcc8725557b9bd52036c12d80a40fa1dfe7ad1dbaefcb"} Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.776454 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-669c8b4d45-5p7x4"] Oct 08 15:31:59 crc kubenswrapper[4945]: I1008 15:31:59.784710 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-669c8b4d45-5p7x4"] Oct 08 15:32:00 crc kubenswrapper[4945]: I1008 15:32:00.033943 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b088ba68-7804-4b57-a1fc-5442e6bf1ef3" path="/var/lib/kubelet/pods/b088ba68-7804-4b57-a1fc-5442e6bf1ef3/volumes" Oct 08 15:32:00 crc kubenswrapper[4945]: I1008 15:32:00.771215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2481ce07-e2b2-4afb-b5f1-027bfefd0f38","Type":"ContainerStarted","Data":"7e947c42db08811c5f07c0d5c81d80376a93939789814059a11b707a33d120e8"} Oct 08 15:32:00 crc kubenswrapper[4945]: I1008 15:32:00.773965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec56f6d2-9c69-47f1-b834-169de890c512","Type":"ContainerStarted","Data":"de13a17be1ddba6e92ab07e7c7fda48d8d412d56c677f0cfff1aa52301f80782"} Oct 08 15:32:00 crc kubenswrapper[4945]: I1008 15:32:00.776150 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"420a6f35-1c79-416f-a7f4-8b9f46e337bc","Type":"ContainerStarted","Data":"83cc5f71b5c95ee829703f3c82db4ea1d047807174a0116d2d6507248426d6a6"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.828147 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cwwpc" event={"ID":"1b70322d-ae60-46b4-828b-1f4a01cae6b6","Type":"ContainerStarted","Data":"ef7bb49857d29e1aa391456b3865cdc1b9516504909f01ad2e80e001859d574e"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.829232 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-cwwpc" Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.831500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4971563b-eba1-4306-a525-e088a5b3f754","Type":"ContainerStarted","Data":"0f074551af9386ef7057a1d44db3e22a5af2d5afa54c054273ef5b22e9eb855c"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.831753 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.835609 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ceb183ce-9008-4a1a-8b6b-26d47b2b246d","Type":"ContainerStarted","Data":"6ede1ba5c5050a8471a99df7a2ac4abca58118ca1ceb637f7d8a1fb9503ddb37"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.835730 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.838219 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1d5103f7-4fab-43a6-8989-1b505ac5d8d0","Type":"ContainerStarted","Data":"877c009856b48f500a2c9e9f6f7cb2092057f4cf5e8a02f0dc006e6980e1a477"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.841168 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" event={"ID":"8316db8b-1cdf-4494-a07d-2552ff4e72e4","Type":"ContainerStarted","Data":"b319eee9135b55b3e591d820172d4813c5b7b2cb187efa10721c292850a57e9e"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.841552 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.844941 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec9d8fdb-5e82-4cb6-811b-053a28588453","Type":"ContainerStarted","Data":"a0cf9d449bb95270176b0b90e00f3b5b23170adb00500a4eb25dc513cc914715"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.848447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" event={"ID":"098407b4-11e2-40a3-b3d1-b47b9d9ca618","Type":"ContainerStarted","Data":"185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.848572 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.851386 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-67zmn" event={"ID":"7fa35d9f-678d-4820-9be9-7db07c5ffd07","Type":"ContainerStarted","Data":"7f1e2bed587442a43ff4a54f1a069cb2486246601000e16a22a8577a87ae14c3"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.854783 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e","Type":"ContainerStarted","Data":"0fa2ef6a70053492713c1f64203ca92929f906e64bfacda033ef7fc5ff49a6f2"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.857614 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c4cb5094-ffbc-4fff-bbc4-0a3a35067782","Type":"ContainerStarted","Data":"6c4780077193c80beb40dc146a66a4baeb56908556174d41812d501b163af399"} Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.867659 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-cwwpc" podStartSLOduration=14.598149235 podStartE2EDuration="19.867647603s" podCreationTimestamp="2025-10-08 15:31:46 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.927280362 +0000 UTC m=+1128.281195263" lastFinishedPulling="2025-10-08 15:32:04.19677872 +0000 UTC m=+1133.550693631" observedRunningTime="2025-10-08 15:32:05.862366839 +0000 UTC m=+1135.216281740" watchObservedRunningTime="2025-10-08 15:32:05.867647603 +0000 UTC m=+1135.221562504" Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.955652 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.745569684 podStartE2EDuration="23.955626485s" podCreationTimestamp="2025-10-08 15:31:42 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.967444116 +0000 UTC m=+1128.321359017" lastFinishedPulling="2025-10-08 15:32:05.177500917 +0000 UTC m=+1134.531415818" observedRunningTime="2025-10-08 15:32:05.952161988 +0000 UTC m=+1135.306076889" watchObservedRunningTime="2025-10-08 15:32:05.955626485 +0000 UTC m=+1135.309541396" Oct 08 15:32:05 crc kubenswrapper[4945]: I1008 15:32:05.978979 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" podStartSLOduration=30.978957655 podStartE2EDuration="30.978957655s" podCreationTimestamp="2025-10-08 15:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:05.974263316 +0000 UTC m=+1135.328178247" watchObservedRunningTime="2025-10-08 15:32:05.978957655 +0000 UTC m=+1135.332872556" Oct 08 15:32:06 crc kubenswrapper[4945]: I1008 15:32:06.001062 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" podStartSLOduration=31.001044643 podStartE2EDuration="31.001044643s" podCreationTimestamp="2025-10-08 15:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:05.996280232 +0000 UTC m=+1135.350195133" watchObservedRunningTime="2025-10-08 15:32:06.001044643 +0000 UTC m=+1135.354959544" Oct 08 15:32:06 crc kubenswrapper[4945]: I1008 15:32:06.017417 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.299611276 podStartE2EDuration="25.017401286s" podCreationTimestamp="2025-10-08 15:31:41 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.952403616 +0000 UTC m=+1128.306318517" lastFinishedPulling="2025-10-08 15:32:03.670193626 +0000 UTC m=+1133.024108527" observedRunningTime="2025-10-08 15:32:06.011683911 +0000 UTC m=+1135.365598812" watchObservedRunningTime="2025-10-08 15:32:06.017401286 +0000 UTC m=+1135.371316187" Oct 08 15:32:06 crc kubenswrapper[4945]: I1008 15:32:06.867964 4945 generic.go:334] "Generic (PLEG): container finished" podID="7fa35d9f-678d-4820-9be9-7db07c5ffd07" containerID="7f1e2bed587442a43ff4a54f1a069cb2486246601000e16a22a8577a87ae14c3" exitCode=0 Oct 08 15:32:06 crc kubenswrapper[4945]: I1008 15:32:06.868008 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-67zmn" event={"ID":"7fa35d9f-678d-4820-9be9-7db07c5ffd07","Type":"ContainerDied","Data":"7f1e2bed587442a43ff4a54f1a069cb2486246601000e16a22a8577a87ae14c3"} Oct 08 15:32:08 crc kubenswrapper[4945]: I1008 15:32:08.892437 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-67zmn" event={"ID":"7fa35d9f-678d-4820-9be9-7db07c5ffd07","Type":"ContainerStarted","Data":"88390d282d58b733a90a5f366f3bd9210d1787d50074e1278eb6970e4c0ea8ca"} Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.915756 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c4cb5094-ffbc-4fff-bbc4-0a3a35067782","Type":"ContainerStarted","Data":"ae2a4c1e0594b1184782310047a573e1db7924f53738316770f96faf54797331"} Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.919018 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec9d8fdb-5e82-4cb6-811b-053a28588453","Type":"ContainerStarted","Data":"0aacb7b6a44998a8cc0148a6002d2be69519972557d6c14cadc6a9603ecb548c"} Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.922579 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-67zmn" event={"ID":"7fa35d9f-678d-4820-9be9-7db07c5ffd07","Type":"ContainerStarted","Data":"35ba31b10e7fd441e9fbd6e2e81d1ae2ceb1d07ad10843aabf2754f1d315ccb3"} Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.922891 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.923007 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.985312 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.989889 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.800469262 podStartE2EDuration="21.9898669s" podCreationTimestamp="2025-10-08 15:31:49 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.957943786 +0000 UTC m=+1128.311858687" lastFinishedPulling="2025-10-08 15:32:10.147341414 +0000 UTC m=+1139.501256325" observedRunningTime="2025-10-08 15:32:10.984607907 +0000 UTC m=+1140.338522898" watchObservedRunningTime="2025-10-08 15:32:10.9898669 +0000 UTC m=+1140.343781801" Oct 08 15:32:10 crc kubenswrapper[4945]: I1008 15:32:10.992921 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=13.832132761 podStartE2EDuration="24.992907177s" podCreationTimestamp="2025-10-08 15:31:46 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.96759785 +0000 UTC m=+1128.321512761" lastFinishedPulling="2025-10-08 15:32:10.128372276 +0000 UTC m=+1139.482287177" observedRunningTime="2025-10-08 15:32:10.958322513 +0000 UTC m=+1140.312237434" watchObservedRunningTime="2025-10-08 15:32:10.992907177 +0000 UTC m=+1140.346822078" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.013468 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-67zmn" podStartSLOduration=19.940699958 podStartE2EDuration="25.013448846s" podCreationTimestamp="2025-10-08 15:31:46 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.930644096 +0000 UTC m=+1128.284559037" lastFinishedPulling="2025-10-08 15:32:04.003392984 +0000 UTC m=+1133.357307925" observedRunningTime="2025-10-08 15:32:11.002494359 +0000 UTC m=+1140.356409280" watchObservedRunningTime="2025-10-08 15:32:11.013448846 +0000 UTC m=+1140.367363747" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.297041 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.352335 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.389283 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.408777 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78bdd78b69-hpwh4"] Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.773784 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.817701 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.934557 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" podUID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerName="dnsmasq-dns" containerID="cri-o://185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a" gracePeriod=10 Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.934665 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 08 15:32:11 crc kubenswrapper[4945]: I1008 15:32:11.982919 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.244756 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-778975bfdc-9k9mh"] Oct 08 15:32:12 crc kubenswrapper[4945]: E1008 15:32:12.245262 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b088ba68-7804-4b57-a1fc-5442e6bf1ef3" containerName="init" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.245279 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b088ba68-7804-4b57-a1fc-5442e6bf1ef3" containerName="init" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.245520 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b088ba68-7804-4b57-a1fc-5442e6bf1ef3" containerName="init" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.246559 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.253890 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.263591 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778975bfdc-9k9mh"] Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.305484 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.341876 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-x7dfs"] Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.342940 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.346831 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.348959 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv4f4\" (UniqueName: \"kubernetes.io/projected/7220fdcd-a064-4b64-a76f-3dde37f9859f-kube-api-access-nv4f4\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.348997 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-ovsdbserver-nb\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.349029 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-config\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.349149 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-dns-svc\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.357353 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-x7dfs"] Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.445153 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.450889 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0817bf-174f-4d4f-b155-041259384465-combined-ca-bundle\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.450989 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0817bf-174f-4d4f-b155-041259384465-config\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451051 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2c0817bf-174f-4d4f-b155-041259384465-ovs-rundir\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx2th\" (UniqueName: \"kubernetes.io/projected/2c0817bf-174f-4d4f-b155-041259384465-kube-api-access-nx2th\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451264 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-dns-svc\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451305 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0817bf-174f-4d4f-b155-041259384465-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451353 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv4f4\" (UniqueName: \"kubernetes.io/projected/7220fdcd-a064-4b64-a76f-3dde37f9859f-kube-api-access-nv4f4\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-ovsdbserver-nb\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451425 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-config\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.451504 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2c0817bf-174f-4d4f-b155-041259384465-ovn-rundir\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.452727 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-dns-svc\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.453412 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-config\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.453976 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-ovsdbserver-nb\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.481321 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv4f4\" (UniqueName: \"kubernetes.io/projected/7220fdcd-a064-4b64-a76f-3dde37f9859f-kube-api-access-nv4f4\") pod \"dnsmasq-dns-778975bfdc-9k9mh\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.536921 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.554590 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0817bf-174f-4d4f-b155-041259384465-combined-ca-bundle\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.554653 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0817bf-174f-4d4f-b155-041259384465-config\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.554704 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2c0817bf-174f-4d4f-b155-041259384465-ovs-rundir\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.554725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx2th\" (UniqueName: \"kubernetes.io/projected/2c0817bf-174f-4d4f-b155-041259384465-kube-api-access-nx2th\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.554758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0817bf-174f-4d4f-b155-041259384465-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.554845 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2c0817bf-174f-4d4f-b155-041259384465-ovn-rundir\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.555164 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2c0817bf-174f-4d4f-b155-041259384465-ovn-rundir\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.555240 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2c0817bf-174f-4d4f-b155-041259384465-ovs-rundir\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.555969 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0817bf-174f-4d4f-b155-041259384465-config\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.560018 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0817bf-174f-4d4f-b155-041259384465-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.568890 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0817bf-174f-4d4f-b155-041259384465-combined-ca-bundle\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.581267 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx2th\" (UniqueName: \"kubernetes.io/projected/2c0817bf-174f-4d4f-b155-041259384465-kube-api-access-nx2th\") pod \"ovn-controller-metrics-x7dfs\" (UID: \"2c0817bf-174f-4d4f-b155-041259384465\") " pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.606507 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.656180 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-dns-svc\") pod \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.660682 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9gn8\" (UniqueName: \"kubernetes.io/projected/098407b4-11e2-40a3-b3d1-b47b9d9ca618-kube-api-access-c9gn8\") pod \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.660747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-config\") pod \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\" (UID: \"098407b4-11e2-40a3-b3d1-b47b9d9ca618\") " Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.675009 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-x7dfs" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.682057 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778975bfdc-9k9mh"] Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.698821 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8545c9dd95-rz987"] Oct 08 15:32:12 crc kubenswrapper[4945]: E1008 15:32:12.699213 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerName="init" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.699231 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerName="init" Oct 08 15:32:12 crc kubenswrapper[4945]: E1008 15:32:12.699282 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerName="dnsmasq-dns" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.699290 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerName="dnsmasq-dns" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.699610 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerName="dnsmasq-dns" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.700554 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.702522 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.706042 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8545c9dd95-rz987"] Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.763856 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-sb\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.765768 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snmkj\" (UniqueName: \"kubernetes.io/projected/420d933e-4371-4b44-8316-a1f757614bd4-kube-api-access-snmkj\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.765872 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-config\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.766077 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-nb\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.766240 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-dns-svc\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.868040 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-sb\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.868110 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snmkj\" (UniqueName: \"kubernetes.io/projected/420d933e-4371-4b44-8316-a1f757614bd4-kube-api-access-snmkj\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.868146 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-config\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.868194 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-nb\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.868231 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-dns-svc\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.869210 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-sb\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.869247 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-config\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.869525 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-dns-svc\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.869620 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-nb\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.918492 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/098407b4-11e2-40a3-b3d1-b47b9d9ca618-kube-api-access-c9gn8" (OuterVolumeSpecName: "kube-api-access-c9gn8") pod "098407b4-11e2-40a3-b3d1-b47b9d9ca618" (UID: "098407b4-11e2-40a3-b3d1-b47b9d9ca618"). InnerVolumeSpecName "kube-api-access-c9gn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.919334 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snmkj\" (UniqueName: \"kubernetes.io/projected/420d933e-4371-4b44-8316-a1f757614bd4-kube-api-access-snmkj\") pod \"dnsmasq-dns-8545c9dd95-rz987\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.948500 4945 generic.go:334] "Generic (PLEG): container finished" podID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" containerID="185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a" exitCode=0 Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.948617 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" event={"ID":"098407b4-11e2-40a3-b3d1-b47b9d9ca618","Type":"ContainerDied","Data":"185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a"} Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.948650 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" event={"ID":"098407b4-11e2-40a3-b3d1-b47b9d9ca618","Type":"ContainerDied","Data":"0823821a5ccc0288c8af602dac6ec2b20bab8bdbe24ab062930c12f181cef67c"} Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.948700 4945 scope.go:117] "RemoveContainer" containerID="185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.948901 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bdd78b69-hpwh4" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.965948 4945 generic.go:334] "Generic (PLEG): container finished" podID="88f0eb1e-a2bc-4b2f-b14d-8cef0261334e" containerID="0fa2ef6a70053492713c1f64203ca92929f906e64bfacda033ef7fc5ff49a6f2" exitCode=0 Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.966147 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e","Type":"ContainerDied","Data":"0fa2ef6a70053492713c1f64203ca92929f906e64bfacda033ef7fc5ff49a6f2"} Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.970720 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9gn8\" (UniqueName: \"kubernetes.io/projected/098407b4-11e2-40a3-b3d1-b47b9d9ca618-kube-api-access-c9gn8\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.978558 4945 generic.go:334] "Generic (PLEG): container finished" podID="1d5103f7-4fab-43a6-8989-1b505ac5d8d0" containerID="877c009856b48f500a2c9e9f6f7cb2092057f4cf5e8a02f0dc006e6980e1a477" exitCode=0 Oct 08 15:32:12 crc kubenswrapper[4945]: I1008 15:32:12.979684 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1d5103f7-4fab-43a6-8989-1b505ac5d8d0","Type":"ContainerDied","Data":"877c009856b48f500a2c9e9f6f7cb2092057f4cf5e8a02f0dc006e6980e1a477"} Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.025630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.066553 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-config" (OuterVolumeSpecName: "config") pod "098407b4-11e2-40a3-b3d1-b47b9d9ca618" (UID: "098407b4-11e2-40a3-b3d1-b47b9d9ca618"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.079663 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.084849 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "098407b4-11e2-40a3-b3d1-b47b9d9ca618" (UID: "098407b4-11e2-40a3-b3d1-b47b9d9ca618"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.089164 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.181728 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/098407b4-11e2-40a3-b3d1-b47b9d9ca618-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.205836 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8545c9dd95-rz987"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.226219 4945 scope.go:117] "RemoveContainer" containerID="275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.245067 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.285330 4945 scope.go:117] "RemoveContainer" containerID="185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a" Oct 08 15:32:13 crc kubenswrapper[4945]: E1008 15:32:13.292866 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a\": container with ID starting with 185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a not found: ID does not exist" containerID="185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.292906 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a"} err="failed to get container status \"185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a\": rpc error: code = NotFound desc = could not find container \"185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a\": container with ID starting with 185dafb31ad338411eb12647c09432c0eac31a507cfafe4b14b0e75751c5847a not found: ID does not exist" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.292929 4945 scope.go:117] "RemoveContainer" containerID="275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.295464 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b89b7f545-vmj78"] Oct 08 15:32:13 crc kubenswrapper[4945]: E1008 15:32:13.300889 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235\": container with ID starting with 275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235 not found: ID does not exist" containerID="275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.300930 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235"} err="failed to get container status \"275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235\": rpc error: code = NotFound desc = could not find container \"275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235\": container with ID starting with 275a75dcbf3960d6869b85282990335d5eb5311866328f2651b718eec0821235 not found: ID does not exist" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.314499 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.333739 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b89b7f545-vmj78"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.489440 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78bdd78b69-hpwh4"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.497228 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78bdd78b69-hpwh4"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.525527 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-sb\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.525595 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-dns-svc\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.525638 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sqsv\" (UniqueName: \"kubernetes.io/projected/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-kube-api-access-2sqsv\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.525701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-config\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.525720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-nb\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.561192 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778975bfdc-9k9mh"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.587530 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.588799 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.594489 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-tc877" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.594694 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.595264 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.595377 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.602290 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636442 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-config\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636481 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-nb\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636510 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636530 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5p4w\" (UniqueName: \"kubernetes.io/projected/f9854a42-6052-4425-a3a8-190bd5533f75-kube-api-access-v5p4w\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636547 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-sb\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636564 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-dns-svc\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636625 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9854a42-6052-4425-a3a8-190bd5533f75-config\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636654 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9854a42-6052-4425-a3a8-190bd5533f75-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636672 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sqsv\" (UniqueName: \"kubernetes.io/projected/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-kube-api-access-2sqsv\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636694 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9854a42-6052-4425-a3a8-190bd5533f75-scripts\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.636719 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.637532 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-config\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.637637 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-nb\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.640824 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-dns-svc\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.643798 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-sb\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.673854 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sqsv\" (UniqueName: \"kubernetes.io/projected/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-kube-api-access-2sqsv\") pod \"dnsmasq-dns-7b89b7f545-vmj78\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.703297 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.738203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.738248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5p4w\" (UniqueName: \"kubernetes.io/projected/f9854a42-6052-4425-a3a8-190bd5533f75-kube-api-access-v5p4w\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.738274 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.738344 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9854a42-6052-4425-a3a8-190bd5533f75-config\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.738383 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9854a42-6052-4425-a3a8-190bd5533f75-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.738416 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9854a42-6052-4425-a3a8-190bd5533f75-scripts\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.738455 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.739177 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9854a42-6052-4425-a3a8-190bd5533f75-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.739504 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9854a42-6052-4425-a3a8-190bd5533f75-config\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.740742 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9854a42-6052-4425-a3a8-190bd5533f75-scripts\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.745713 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.746081 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.754855 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9854a42-6052-4425-a3a8-190bd5533f75-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.757030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5p4w\" (UniqueName: \"kubernetes.io/projected/f9854a42-6052-4425-a3a8-190bd5533f75-kube-api-access-v5p4w\") pod \"ovn-northd-0\" (UID: \"f9854a42-6052-4425-a3a8-190bd5533f75\") " pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.805226 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8545c9dd95-rz987"] Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.818359 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-x7dfs"] Oct 08 15:32:13 crc kubenswrapper[4945]: W1008 15:32:13.832915 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c0817bf_174f_4d4f_b155_041259384465.slice/crio-0c4d41d887aee1b8d264437598f2e3631e571fa09897eb5fa64e52d8f75e0178 WatchSource:0}: Error finding container 0c4d41d887aee1b8d264437598f2e3631e571fa09897eb5fa64e52d8f75e0178: Status 404 returned error can't find the container with id 0c4d41d887aee1b8d264437598f2e3631e571fa09897eb5fa64e52d8f75e0178 Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.970820 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.988684 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-x7dfs" event={"ID":"2c0817bf-174f-4d4f-b155-041259384465","Type":"ContainerStarted","Data":"0c4d41d887aee1b8d264437598f2e3631e571fa09897eb5fa64e52d8f75e0178"} Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.990591 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8545c9dd95-rz987" event={"ID":"420d933e-4371-4b44-8316-a1f757614bd4","Type":"ContainerStarted","Data":"3004ca5b75279d3cca7ca664b57ae53d3fa23506952445e19f6b8dcdd4969cb1"} Oct 08 15:32:13 crc kubenswrapper[4945]: I1008 15:32:13.991584 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" event={"ID":"7220fdcd-a064-4b64-a76f-3dde37f9859f","Type":"ContainerStarted","Data":"35d62832baea406943c15e45909b021aaf63e0f1c5407aeef5f413c34041de94"} Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.083865 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="098407b4-11e2-40a3-b3d1-b47b9d9ca618" path="/var/lib/kubelet/pods/098407b4-11e2-40a3-b3d1-b47b9d9ca618/volumes" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.215670 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b89b7f545-vmj78"] Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.416004 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.428203 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.430915 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.431222 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-kvzhj" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.431371 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.431727 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.441493 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.470017 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-nqt9s"] Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.471300 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.476780 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.477055 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.477280 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.494078 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nqt9s"] Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.501806 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.575880 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e8011e91-9053-4a34-a58a-7d046d51536f-cache\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.575968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdn7q\" (UniqueName: \"kubernetes.io/projected/77c71453-5d71-4c06-b1d8-06561a1b2778-kube-api-access-pdn7q\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.576011 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-ring-data-devices\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.576034 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-scripts\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.576058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-swiftconf\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.576166 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e8011e91-9053-4a34-a58a-7d046d51536f-lock\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.576226 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77c71453-5d71-4c06-b1d8-06561a1b2778-etc-swift\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.576255 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.576492 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-dispersionconf\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.577021 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpvwf\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-kube-api-access-gpvwf\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.577068 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-combined-ca-bundle\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.577096 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678211 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77c71453-5d71-4c06-b1d8-06561a1b2778-etc-swift\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678392 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-dispersionconf\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678433 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpvwf\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-kube-api-access-gpvwf\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-combined-ca-bundle\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678491 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678528 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e8011e91-9053-4a34-a58a-7d046d51536f-cache\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678579 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdn7q\" (UniqueName: \"kubernetes.io/projected/77c71453-5d71-4c06-b1d8-06561a1b2778-kube-api-access-pdn7q\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-ring-data-devices\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678621 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-scripts\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678644 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-swiftconf\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678658 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e8011e91-9053-4a34-a58a-7d046d51536f-lock\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77c71453-5d71-4c06-b1d8-06561a1b2778-etc-swift\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.678776 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.679233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e8011e91-9053-4a34-a58a-7d046d51536f-lock\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: E1008 15:32:14.679404 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 15:32:14 crc kubenswrapper[4945]: E1008 15:32:14.679427 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 15:32:14 crc kubenswrapper[4945]: E1008 15:32:14.679464 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift podName:e8011e91-9053-4a34-a58a-7d046d51536f nodeName:}" failed. No retries permitted until 2025-10-08 15:32:15.179450644 +0000 UTC m=+1144.533365545 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift") pod "swift-storage-0" (UID: "e8011e91-9053-4a34-a58a-7d046d51536f") : configmap "swift-ring-files" not found Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.679881 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-ring-data-devices\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.680091 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-scripts\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.680349 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e8011e91-9053-4a34-a58a-7d046d51536f-cache\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.684403 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-dispersionconf\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.684563 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-swiftconf\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.684907 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-combined-ca-bundle\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.697346 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpvwf\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-kube-api-access-gpvwf\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.699014 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdn7q\" (UniqueName: \"kubernetes.io/projected/77c71453-5d71-4c06-b1d8-06561a1b2778-kube-api-access-pdn7q\") pod \"swift-ring-rebalance-nqt9s\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.703698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:14 crc kubenswrapper[4945]: I1008 15:32:14.845621 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.001673 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerStarted","Data":"a07652c59d10c5c9ff261dfa0a5d05794c80ce3a34b1b4a4f5d2001b7fa1ab53"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.012598 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1d5103f7-4fab-43a6-8989-1b505ac5d8d0","Type":"ContainerStarted","Data":"a2fc03e5db6780eb96606be948f3fb1bdfdba0f9917c269e7045714dcca3ac11"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.023404 4945 generic.go:334] "Generic (PLEG): container finished" podID="420d933e-4371-4b44-8316-a1f757614bd4" containerID="32da93bc946f2aedc43c97e130e411eea2959c9b6a3ee5e383a16ab79668adb3" exitCode=0 Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.023538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8545c9dd95-rz987" event={"ID":"420d933e-4371-4b44-8316-a1f757614bd4","Type":"ContainerDied","Data":"32da93bc946f2aedc43c97e130e411eea2959c9b6a3ee5e383a16ab79668adb3"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.034036 4945 generic.go:334] "Generic (PLEG): container finished" podID="7220fdcd-a064-4b64-a76f-3dde37f9859f" containerID="9164820f1a5add6a4858cf7f2dcfacd0d7deb97ca2269f8ae9a6a367dea8301c" exitCode=0 Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.034204 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" event={"ID":"7220fdcd-a064-4b64-a76f-3dde37f9859f","Type":"ContainerDied","Data":"9164820f1a5add6a4858cf7f2dcfacd0d7deb97ca2269f8ae9a6a367dea8301c"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.060585 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-x7dfs" event={"ID":"2c0817bf-174f-4d4f-b155-041259384465","Type":"ContainerStarted","Data":"7253624eb236e8c96eef6f691f5552617f2e4b9aa8f1b72015d086f4db232b75"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.079663 4945 generic.go:334] "Generic (PLEG): container finished" podID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerID="9ac990b826437d87d72086f7681687339824a7e9e25cb806c64d7af7353331c9" exitCode=0 Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.079752 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" event={"ID":"acf8ac52-ea0c-4ea9-9a19-616edcd554eb","Type":"ContainerDied","Data":"9ac990b826437d87d72086f7681687339824a7e9e25cb806c64d7af7353331c9"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.079777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" event={"ID":"acf8ac52-ea0c-4ea9-9a19-616edcd554eb","Type":"ContainerStarted","Data":"2de36999dfb27382d489c426a443d5806a5e1ffc99eeae999c8b8ac3f1170bd7"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.093212 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f9854a42-6052-4425-a3a8-190bd5533f75","Type":"ContainerStarted","Data":"8bd261cc785aedf52dba2c318ce66bef3540f935bff46b5acfd8bfea08f6c812"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.094063 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=30.951199259 podStartE2EDuration="36.094049088s" podCreationTimestamp="2025-10-08 15:31:39 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.944366893 +0000 UTC m=+1128.298281794" lastFinishedPulling="2025-10-08 15:32:04.087216712 +0000 UTC m=+1133.441131623" observedRunningTime="2025-10-08 15:32:15.093553586 +0000 UTC m=+1144.447468487" watchObservedRunningTime="2025-10-08 15:32:15.094049088 +0000 UTC m=+1144.447963989" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.098933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"88f0eb1e-a2bc-4b2f-b14d-8cef0261334e","Type":"ContainerStarted","Data":"132e349c8862cb021f97f3edb18702da64317bda667f39d3aa17e924aa1e70e7"} Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.112245 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-x7dfs" podStartSLOduration=3.112226147 podStartE2EDuration="3.112226147s" podCreationTimestamp="2025-10-08 15:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:15.11193728 +0000 UTC m=+1144.465852181" watchObservedRunningTime="2025-10-08 15:32:15.112226147 +0000 UTC m=+1144.466141048" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.186528 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:15 crc kubenswrapper[4945]: E1008 15:32:15.188860 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 15:32:15 crc kubenswrapper[4945]: E1008 15:32:15.188876 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 15:32:15 crc kubenswrapper[4945]: E1008 15:32:15.188910 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift podName:e8011e91-9053-4a34-a58a-7d046d51536f nodeName:}" failed. No retries permitted until 2025-10-08 15:32:16.188896974 +0000 UTC m=+1145.542811875 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift") pod "swift-storage-0" (UID: "e8011e91-9053-4a34-a58a-7d046d51536f") : configmap "swift-ring-files" not found Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.190461 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=30.806132723 podStartE2EDuration="36.190438623s" podCreationTimestamp="2025-10-08 15:31:39 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.960383277 +0000 UTC m=+1128.314298178" lastFinishedPulling="2025-10-08 15:32:04.344689177 +0000 UTC m=+1133.698604078" observedRunningTime="2025-10-08 15:32:15.182856182 +0000 UTC m=+1144.536771093" watchObservedRunningTime="2025-10-08 15:32:15.190438623 +0000 UTC m=+1144.544353524" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.314961 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nqt9s"] Oct 08 15:32:15 crc kubenswrapper[4945]: W1008 15:32:15.569339 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77c71453_5d71_4c06_b1d8_06561a1b2778.slice/crio-f7cba48e50550e093156c00bde8eef97bcbd8c3ad0033cae5f5a436b11dd24c0 WatchSource:0}: Error finding container f7cba48e50550e093156c00bde8eef97bcbd8c3ad0033cae5f5a436b11dd24c0: Status 404 returned error can't find the container with id f7cba48e50550e093156c00bde8eef97bcbd8c3ad0033cae5f5a436b11dd24c0 Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.763592 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.805144 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.900953 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-config\") pod \"420d933e-4371-4b44-8316-a1f757614bd4\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.900988 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-config\") pod \"7220fdcd-a064-4b64-a76f-3dde37f9859f\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.901013 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-nb\") pod \"420d933e-4371-4b44-8316-a1f757614bd4\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.901033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-ovsdbserver-nb\") pod \"7220fdcd-a064-4b64-a76f-3dde37f9859f\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.901058 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-dns-svc\") pod \"7220fdcd-a064-4b64-a76f-3dde37f9859f\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.901083 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snmkj\" (UniqueName: \"kubernetes.io/projected/420d933e-4371-4b44-8316-a1f757614bd4-kube-api-access-snmkj\") pod \"420d933e-4371-4b44-8316-a1f757614bd4\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.901117 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-sb\") pod \"420d933e-4371-4b44-8316-a1f757614bd4\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.901251 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv4f4\" (UniqueName: \"kubernetes.io/projected/7220fdcd-a064-4b64-a76f-3dde37f9859f-kube-api-access-nv4f4\") pod \"7220fdcd-a064-4b64-a76f-3dde37f9859f\" (UID: \"7220fdcd-a064-4b64-a76f-3dde37f9859f\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.901298 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-dns-svc\") pod \"420d933e-4371-4b44-8316-a1f757614bd4\" (UID: \"420d933e-4371-4b44-8316-a1f757614bd4\") " Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.904672 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/420d933e-4371-4b44-8316-a1f757614bd4-kube-api-access-snmkj" (OuterVolumeSpecName: "kube-api-access-snmkj") pod "420d933e-4371-4b44-8316-a1f757614bd4" (UID: "420d933e-4371-4b44-8316-a1f757614bd4"). InnerVolumeSpecName "kube-api-access-snmkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.907024 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7220fdcd-a064-4b64-a76f-3dde37f9859f-kube-api-access-nv4f4" (OuterVolumeSpecName: "kube-api-access-nv4f4") pod "7220fdcd-a064-4b64-a76f-3dde37f9859f" (UID: "7220fdcd-a064-4b64-a76f-3dde37f9859f"). InnerVolumeSpecName "kube-api-access-nv4f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.924806 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-config" (OuterVolumeSpecName: "config") pod "7220fdcd-a064-4b64-a76f-3dde37f9859f" (UID: "7220fdcd-a064-4b64-a76f-3dde37f9859f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.927548 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7220fdcd-a064-4b64-a76f-3dde37f9859f" (UID: "7220fdcd-a064-4b64-a76f-3dde37f9859f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.927514 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-config" (OuterVolumeSpecName: "config") pod "420d933e-4371-4b44-8316-a1f757614bd4" (UID: "420d933e-4371-4b44-8316-a1f757614bd4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.929433 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "420d933e-4371-4b44-8316-a1f757614bd4" (UID: "420d933e-4371-4b44-8316-a1f757614bd4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.930619 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "420d933e-4371-4b44-8316-a1f757614bd4" (UID: "420d933e-4371-4b44-8316-a1f757614bd4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.932236 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "420d933e-4371-4b44-8316-a1f757614bd4" (UID: "420d933e-4371-4b44-8316-a1f757614bd4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:15 crc kubenswrapper[4945]: I1008 15:32:15.932540 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7220fdcd-a064-4b64-a76f-3dde37f9859f" (UID: "7220fdcd-a064-4b64-a76f-3dde37f9859f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003227 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003263 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003277 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003289 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7220fdcd-a064-4b64-a76f-3dde37f9859f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003301 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snmkj\" (UniqueName: \"kubernetes.io/projected/420d933e-4371-4b44-8316-a1f757614bd4-kube-api-access-snmkj\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003314 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003327 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv4f4\" (UniqueName: \"kubernetes.io/projected/7220fdcd-a064-4b64-a76f-3dde37f9859f-kube-api-access-nv4f4\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003339 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.003348 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420d933e-4371-4b44-8316-a1f757614bd4-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.111807 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" event={"ID":"acf8ac52-ea0c-4ea9-9a19-616edcd554eb","Type":"ContainerStarted","Data":"b7cd398e2b71fc245f28a3235d309f80fb40f4f5747176ebdaf555f8bc1c2cf5"} Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.112211 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.117803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f9854a42-6052-4425-a3a8-190bd5533f75","Type":"ContainerStarted","Data":"28db2f58489bb1933df98045163913cc804df65dcd5bc18926399dbb776bcd88"} Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.117842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f9854a42-6052-4425-a3a8-190bd5533f75","Type":"ContainerStarted","Data":"155c5ce7d1939c15af90cede0c0be300f5d725dcf2e99579fc1d480c5db0cc84"} Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.118088 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.119634 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqt9s" event={"ID":"77c71453-5d71-4c06-b1d8-06561a1b2778","Type":"ContainerStarted","Data":"f7cba48e50550e093156c00bde8eef97bcbd8c3ad0033cae5f5a436b11dd24c0"} Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.121592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" event={"ID":"7220fdcd-a064-4b64-a76f-3dde37f9859f","Type":"ContainerDied","Data":"35d62832baea406943c15e45909b021aaf63e0f1c5407aeef5f413c34041de94"} Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.121622 4945 scope.go:117] "RemoveContainer" containerID="9164820f1a5add6a4858cf7f2dcfacd0d7deb97ca2269f8ae9a6a367dea8301c" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.121647 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778975bfdc-9k9mh" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.124191 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8545c9dd95-rz987" event={"ID":"420d933e-4371-4b44-8316-a1f757614bd4","Type":"ContainerDied","Data":"3004ca5b75279d3cca7ca664b57ae53d3fa23506952445e19f6b8dcdd4969cb1"} Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.124269 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8545c9dd95-rz987" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.146201 4945 scope.go:117] "RemoveContainer" containerID="32da93bc946f2aedc43c97e130e411eea2959c9b6a3ee5e383a16ab79668adb3" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.164684 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" podStartSLOduration=3.164663086 podStartE2EDuration="3.164663086s" podCreationTimestamp="2025-10-08 15:32:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:16.134432722 +0000 UTC m=+1145.488347623" watchObservedRunningTime="2025-10-08 15:32:16.164663086 +0000 UTC m=+1145.518577987" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.195005 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.090472919 podStartE2EDuration="3.194987673s" podCreationTimestamp="2025-10-08 15:32:13 +0000 UTC" firstStartedPulling="2025-10-08 15:32:14.514112687 +0000 UTC m=+1143.868027588" lastFinishedPulling="2025-10-08 15:32:15.618627441 +0000 UTC m=+1144.972542342" observedRunningTime="2025-10-08 15:32:16.17076405 +0000 UTC m=+1145.524678951" watchObservedRunningTime="2025-10-08 15:32:16.194987673 +0000 UTC m=+1145.548902574" Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.210200 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:16 crc kubenswrapper[4945]: E1008 15:32:16.212292 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 15:32:16 crc kubenswrapper[4945]: E1008 15:32:16.212322 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 15:32:16 crc kubenswrapper[4945]: E1008 15:32:16.212359 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift podName:e8011e91-9053-4a34-a58a-7d046d51536f nodeName:}" failed. No retries permitted until 2025-10-08 15:32:18.212344571 +0000 UTC m=+1147.566259472 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift") pod "swift-storage-0" (UID: "e8011e91-9053-4a34-a58a-7d046d51536f") : configmap "swift-ring-files" not found Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.220515 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8545c9dd95-rz987"] Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.234202 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8545c9dd95-rz987"] Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.248811 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778975bfdc-9k9mh"] Oct 08 15:32:16 crc kubenswrapper[4945]: I1008 15:32:16.257028 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-778975bfdc-9k9mh"] Oct 08 15:32:18 crc kubenswrapper[4945]: I1008 15:32:18.050736 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="420d933e-4371-4b44-8316-a1f757614bd4" path="/var/lib/kubelet/pods/420d933e-4371-4b44-8316-a1f757614bd4/volumes" Oct 08 15:32:18 crc kubenswrapper[4945]: I1008 15:32:18.052736 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7220fdcd-a064-4b64-a76f-3dde37f9859f" path="/var/lib/kubelet/pods/7220fdcd-a064-4b64-a76f-3dde37f9859f/volumes" Oct 08 15:32:18 crc kubenswrapper[4945]: I1008 15:32:18.245571 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:18 crc kubenswrapper[4945]: E1008 15:32:18.245783 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 15:32:18 crc kubenswrapper[4945]: E1008 15:32:18.245796 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 15:32:18 crc kubenswrapper[4945]: E1008 15:32:18.245841 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift podName:e8011e91-9053-4a34-a58a-7d046d51536f nodeName:}" failed. No retries permitted until 2025-10-08 15:32:22.245827764 +0000 UTC m=+1151.599742665 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift") pod "swift-storage-0" (UID: "e8011e91-9053-4a34-a58a-7d046d51536f") : configmap "swift-ring-files" not found Oct 08 15:32:19 crc kubenswrapper[4945]: I1008 15:32:19.184803 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:32:19 crc kubenswrapper[4945]: I1008 15:32:19.184866 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:32:20 crc kubenswrapper[4945]: I1008 15:32:20.165475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqt9s" event={"ID":"77c71453-5d71-4c06-b1d8-06561a1b2778","Type":"ContainerStarted","Data":"dd700ec3d4a3997998a7dc1c9eab3ecbf316202622f314b393275712ceee797c"} Oct 08 15:32:20 crc kubenswrapper[4945]: I1008 15:32:20.192663 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-nqt9s" podStartSLOduration=2.279024336 podStartE2EDuration="6.192642729s" podCreationTimestamp="2025-10-08 15:32:14 +0000 UTC" firstStartedPulling="2025-10-08 15:32:15.614977029 +0000 UTC m=+1144.968891930" lastFinishedPulling="2025-10-08 15:32:19.528595422 +0000 UTC m=+1148.882510323" observedRunningTime="2025-10-08 15:32:20.182573025 +0000 UTC m=+1149.536487936" watchObservedRunningTime="2025-10-08 15:32:20.192642729 +0000 UTC m=+1149.546557650" Oct 08 15:32:20 crc kubenswrapper[4945]: I1008 15:32:20.900985 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 08 15:32:20 crc kubenswrapper[4945]: I1008 15:32:20.901040 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 08 15:32:20 crc kubenswrapper[4945]: I1008 15:32:20.934754 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 08 15:32:20 crc kubenswrapper[4945]: I1008 15:32:20.934812 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 08 15:32:21 crc kubenswrapper[4945]: I1008 15:32:21.626332 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 08 15:32:21 crc kubenswrapper[4945]: I1008 15:32:21.718195 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="88f0eb1e-a2bc-4b2f-b14d-8cef0261334e" containerName="galera" probeResult="failure" output=< Oct 08 15:32:21 crc kubenswrapper[4945]: wsrep_local_state_comment (Joined) differs from Synced Oct 08 15:32:21 crc kubenswrapper[4945]: > Oct 08 15:32:22 crc kubenswrapper[4945]: I1008 15:32:22.193191 4945 generic.go:334] "Generic (PLEG): container finished" podID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerID="a07652c59d10c5c9ff261dfa0a5d05794c80ce3a34b1b4a4f5d2001b7fa1ab53" exitCode=0 Oct 08 15:32:22 crc kubenswrapper[4945]: I1008 15:32:22.193281 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerDied","Data":"a07652c59d10c5c9ff261dfa0a5d05794c80ce3a34b1b4a4f5d2001b7fa1ab53"} Oct 08 15:32:22 crc kubenswrapper[4945]: I1008 15:32:22.328198 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:22 crc kubenswrapper[4945]: E1008 15:32:22.328393 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 15:32:22 crc kubenswrapper[4945]: E1008 15:32:22.328419 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 15:32:22 crc kubenswrapper[4945]: E1008 15:32:22.328821 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift podName:e8011e91-9053-4a34-a58a-7d046d51536f nodeName:}" failed. No retries permitted until 2025-10-08 15:32:30.328805256 +0000 UTC m=+1159.682720157 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift") pod "swift-storage-0" (UID: "e8011e91-9053-4a34-a58a-7d046d51536f") : configmap "swift-ring-files" not found Oct 08 15:32:23 crc kubenswrapper[4945]: I1008 15:32:23.017536 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 08 15:32:23 crc kubenswrapper[4945]: I1008 15:32:23.108182 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 08 15:32:23 crc kubenswrapper[4945]: I1008 15:32:23.705371 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:32:23 crc kubenswrapper[4945]: I1008 15:32:23.770756 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9bb9d48f5-99b2z"] Oct 08 15:32:23 crc kubenswrapper[4945]: I1008 15:32:23.771011 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" podUID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerName="dnsmasq-dns" containerID="cri-o://b319eee9135b55b3e591d820172d4813c5b7b2cb187efa10721c292850a57e9e" gracePeriod=10 Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.217720 4945 generic.go:334] "Generic (PLEG): container finished" podID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerID="b319eee9135b55b3e591d820172d4813c5b7b2cb187efa10721c292850a57e9e" exitCode=0 Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.218338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" event={"ID":"8316db8b-1cdf-4494-a07d-2552ff4e72e4","Type":"ContainerDied","Data":"b319eee9135b55b3e591d820172d4813c5b7b2cb187efa10721c292850a57e9e"} Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.309375 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.466769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-dns-svc\") pod \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.467198 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npwpf\" (UniqueName: \"kubernetes.io/projected/8316db8b-1cdf-4494-a07d-2552ff4e72e4-kube-api-access-npwpf\") pod \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.467221 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-config\") pod \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\" (UID: \"8316db8b-1cdf-4494-a07d-2552ff4e72e4\") " Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.473797 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8316db8b-1cdf-4494-a07d-2552ff4e72e4-kube-api-access-npwpf" (OuterVolumeSpecName: "kube-api-access-npwpf") pod "8316db8b-1cdf-4494-a07d-2552ff4e72e4" (UID: "8316db8b-1cdf-4494-a07d-2552ff4e72e4"). InnerVolumeSpecName "kube-api-access-npwpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.511696 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-config" (OuterVolumeSpecName: "config") pod "8316db8b-1cdf-4494-a07d-2552ff4e72e4" (UID: "8316db8b-1cdf-4494-a07d-2552ff4e72e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.516311 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8316db8b-1cdf-4494-a07d-2552ff4e72e4" (UID: "8316db8b-1cdf-4494-a07d-2552ff4e72e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.569559 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.569602 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npwpf\" (UniqueName: \"kubernetes.io/projected/8316db8b-1cdf-4494-a07d-2552ff4e72e4-kube-api-access-npwpf\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:24 crc kubenswrapper[4945]: I1008 15:32:24.569620 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8316db8b-1cdf-4494-a07d-2552ff4e72e4-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:25 crc kubenswrapper[4945]: I1008 15:32:25.229959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" event={"ID":"8316db8b-1cdf-4494-a07d-2552ff4e72e4","Type":"ContainerDied","Data":"99a710e4c9e8c848e67a9574a282222abd9f688d918e8d64abfa15c679d3e34e"} Oct 08 15:32:25 crc kubenswrapper[4945]: I1008 15:32:25.230019 4945 scope.go:117] "RemoveContainer" containerID="b319eee9135b55b3e591d820172d4813c5b7b2cb187efa10721c292850a57e9e" Oct 08 15:32:25 crc kubenswrapper[4945]: I1008 15:32:25.230041 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bb9d48f5-99b2z" Oct 08 15:32:25 crc kubenswrapper[4945]: I1008 15:32:25.253557 4945 scope.go:117] "RemoveContainer" containerID="d631b88dcaa75516846337ba829f3a4968b32080ee8bbb8b55b9a478e6c82488" Oct 08 15:32:25 crc kubenswrapper[4945]: I1008 15:32:25.272670 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9bb9d48f5-99b2z"] Oct 08 15:32:25 crc kubenswrapper[4945]: I1008 15:32:25.281588 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9bb9d48f5-99b2z"] Oct 08 15:32:26 crc kubenswrapper[4945]: I1008 15:32:26.037284 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" path="/var/lib/kubelet/pods/8316db8b-1cdf-4494-a07d-2552ff4e72e4/volumes" Oct 08 15:32:29 crc kubenswrapper[4945]: I1008 15:32:29.056943 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 08 15:32:29 crc kubenswrapper[4945]: I1008 15:32:29.263255 4945 generic.go:334] "Generic (PLEG): container finished" podID="77c71453-5d71-4c06-b1d8-06561a1b2778" containerID="dd700ec3d4a3997998a7dc1c9eab3ecbf316202622f314b393275712ceee797c" exitCode=0 Oct 08 15:32:29 crc kubenswrapper[4945]: I1008 15:32:29.263303 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqt9s" event={"ID":"77c71453-5d71-4c06-b1d8-06561a1b2778","Type":"ContainerDied","Data":"dd700ec3d4a3997998a7dc1c9eab3ecbf316202622f314b393275712ceee797c"} Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.414749 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.422252 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8011e91-9053-4a34-a58a-7d046d51536f-etc-swift\") pod \"swift-storage-0\" (UID: \"e8011e91-9053-4a34-a58a-7d046d51536f\") " pod="openstack/swift-storage-0" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.438239 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.606925 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.719729 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-dispersionconf\") pod \"77c71453-5d71-4c06-b1d8-06561a1b2778\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.719890 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-combined-ca-bundle\") pod \"77c71453-5d71-4c06-b1d8-06561a1b2778\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.719956 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-scripts\") pod \"77c71453-5d71-4c06-b1d8-06561a1b2778\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.719978 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77c71453-5d71-4c06-b1d8-06561a1b2778-etc-swift\") pod \"77c71453-5d71-4c06-b1d8-06561a1b2778\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.720021 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-ring-data-devices\") pod \"77c71453-5d71-4c06-b1d8-06561a1b2778\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.720040 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-swiftconf\") pod \"77c71453-5d71-4c06-b1d8-06561a1b2778\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.720061 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdn7q\" (UniqueName: \"kubernetes.io/projected/77c71453-5d71-4c06-b1d8-06561a1b2778-kube-api-access-pdn7q\") pod \"77c71453-5d71-4c06-b1d8-06561a1b2778\" (UID: \"77c71453-5d71-4c06-b1d8-06561a1b2778\") " Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.721276 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77c71453-5d71-4c06-b1d8-06561a1b2778-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "77c71453-5d71-4c06-b1d8-06561a1b2778" (UID: "77c71453-5d71-4c06-b1d8-06561a1b2778"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.722001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "77c71453-5d71-4c06-b1d8-06561a1b2778" (UID: "77c71453-5d71-4c06-b1d8-06561a1b2778"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.725007 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77c71453-5d71-4c06-b1d8-06561a1b2778-kube-api-access-pdn7q" (OuterVolumeSpecName: "kube-api-access-pdn7q") pod "77c71453-5d71-4c06-b1d8-06561a1b2778" (UID: "77c71453-5d71-4c06-b1d8-06561a1b2778"). InnerVolumeSpecName "kube-api-access-pdn7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.727481 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "77c71453-5d71-4c06-b1d8-06561a1b2778" (UID: "77c71453-5d71-4c06-b1d8-06561a1b2778"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.743907 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77c71453-5d71-4c06-b1d8-06561a1b2778" (UID: "77c71453-5d71-4c06-b1d8-06561a1b2778"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.744349 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "77c71453-5d71-4c06-b1d8-06561a1b2778" (UID: "77c71453-5d71-4c06-b1d8-06561a1b2778"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.750195 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-scripts" (OuterVolumeSpecName: "scripts") pod "77c71453-5d71-4c06-b1d8-06561a1b2778" (UID: "77c71453-5d71-4c06-b1d8-06561a1b2778"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.822367 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdn7q\" (UniqueName: \"kubernetes.io/projected/77c71453-5d71-4c06-b1d8-06561a1b2778-kube-api-access-pdn7q\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.822398 4945 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.822410 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.822420 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.822429 4945 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77c71453-5d71-4c06-b1d8-06561a1b2778-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.822436 4945 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77c71453-5d71-4c06-b1d8-06561a1b2778-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.822445 4945 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77c71453-5d71-4c06-b1d8-06561a1b2778-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:30 crc kubenswrapper[4945]: I1008 15:32:30.996200 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.016169 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.296258 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"417e3abf1a9abc4327c0166629cebca231c1b5b802a3989938046964546f45b5"} Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.300436 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqt9s" event={"ID":"77c71453-5d71-4c06-b1d8-06561a1b2778","Type":"ContainerDied","Data":"f7cba48e50550e093156c00bde8eef97bcbd8c3ad0033cae5f5a436b11dd24c0"} Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.300474 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7cba48e50550e093156c00bde8eef97bcbd8c3ad0033cae5f5a436b11dd24c0" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.300548 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqt9s" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.304664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerStarted","Data":"e719bddc4017dc7bcdc4d982354494cbd4619fc87e3ca76e4cde77b0e9d7b204"} Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.428838 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-n5pnx"] Oct 08 15:32:31 crc kubenswrapper[4945]: E1008 15:32:31.429442 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7220fdcd-a064-4b64-a76f-3dde37f9859f" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429458 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7220fdcd-a064-4b64-a76f-3dde37f9859f" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: E1008 15:32:31.429488 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429497 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: E1008 15:32:31.429508 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420d933e-4371-4b44-8316-a1f757614bd4" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429517 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="420d933e-4371-4b44-8316-a1f757614bd4" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: E1008 15:32:31.429535 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c71453-5d71-4c06-b1d8-06561a1b2778" containerName="swift-ring-rebalance" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429543 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c71453-5d71-4c06-b1d8-06561a1b2778" containerName="swift-ring-rebalance" Oct 08 15:32:31 crc kubenswrapper[4945]: E1008 15:32:31.429558 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerName="dnsmasq-dns" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429567 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerName="dnsmasq-dns" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429782 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8316db8b-1cdf-4494-a07d-2552ff4e72e4" containerName="dnsmasq-dns" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429796 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7220fdcd-a064-4b64-a76f-3dde37f9859f" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429819 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="420d933e-4371-4b44-8316-a1f757614bd4" containerName="init" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.429840 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="77c71453-5d71-4c06-b1d8-06561a1b2778" containerName="swift-ring-rebalance" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.430535 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n5pnx" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.442045 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-n5pnx"] Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.536681 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86lqz\" (UniqueName: \"kubernetes.io/projected/4cc1f5af-e985-4b88-bb14-2d4cd9555188-kube-api-access-86lqz\") pod \"placement-db-create-n5pnx\" (UID: \"4cc1f5af-e985-4b88-bb14-2d4cd9555188\") " pod="openstack/placement-db-create-n5pnx" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.638303 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86lqz\" (UniqueName: \"kubernetes.io/projected/4cc1f5af-e985-4b88-bb14-2d4cd9555188-kube-api-access-86lqz\") pod \"placement-db-create-n5pnx\" (UID: \"4cc1f5af-e985-4b88-bb14-2d4cd9555188\") " pod="openstack/placement-db-create-n5pnx" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.660102 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86lqz\" (UniqueName: \"kubernetes.io/projected/4cc1f5af-e985-4b88-bb14-2d4cd9555188-kube-api-access-86lqz\") pod \"placement-db-create-n5pnx\" (UID: \"4cc1f5af-e985-4b88-bb14-2d4cd9555188\") " pod="openstack/placement-db-create-n5pnx" Oct 08 15:32:31 crc kubenswrapper[4945]: I1008 15:32:31.743146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n5pnx" Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.180878 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-n5pnx"] Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.198632 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-8lkkh"] Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.200395 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-8lkkh" Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.207647 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-8lkkh"] Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.265084 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69tqb\" (UniqueName: \"kubernetes.io/projected/817d5ad4-17fa-4845-9a91-c96e271d4d40-kube-api-access-69tqb\") pod \"watcher-db-create-8lkkh\" (UID: \"817d5ad4-17fa-4845-9a91-c96e271d4d40\") " pod="openstack/watcher-db-create-8lkkh" Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.324874 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec56f6d2-9c69-47f1-b834-169de890c512" containerID="de13a17be1ddba6e92ab07e7c7fda48d8d412d56c677f0cfff1aa52301f80782" exitCode=0 Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.324949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec56f6d2-9c69-47f1-b834-169de890c512","Type":"ContainerDied","Data":"de13a17be1ddba6e92ab07e7c7fda48d8d412d56c677f0cfff1aa52301f80782"} Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.327557 4945 generic.go:334] "Generic (PLEG): container finished" podID="420a6f35-1c79-416f-a7f4-8b9f46e337bc" containerID="83cc5f71b5c95ee829703f3c82db4ea1d047807174a0116d2d6507248426d6a6" exitCode=0 Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.327628 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"420a6f35-1c79-416f-a7f4-8b9f46e337bc","Type":"ContainerDied","Data":"83cc5f71b5c95ee829703f3c82db4ea1d047807174a0116d2d6507248426d6a6"} Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.334887 4945 generic.go:334] "Generic (PLEG): container finished" podID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerID="7e947c42db08811c5f07c0d5c81d80376a93939789814059a11b707a33d120e8" exitCode=0 Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.334965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2481ce07-e2b2-4afb-b5f1-027bfefd0f38","Type":"ContainerDied","Data":"7e947c42db08811c5f07c0d5c81d80376a93939789814059a11b707a33d120e8"} Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.344249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n5pnx" event={"ID":"4cc1f5af-e985-4b88-bb14-2d4cd9555188","Type":"ContainerStarted","Data":"b1f812e0b2f91c9c4590008873d9e6b656bfa249a342ff3a9d6d46a74c4c2975"} Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.366474 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69tqb\" (UniqueName: \"kubernetes.io/projected/817d5ad4-17fa-4845-9a91-c96e271d4d40-kube-api-access-69tqb\") pod \"watcher-db-create-8lkkh\" (UID: \"817d5ad4-17fa-4845-9a91-c96e271d4d40\") " pod="openstack/watcher-db-create-8lkkh" Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.399368 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69tqb\" (UniqueName: \"kubernetes.io/projected/817d5ad4-17fa-4845-9a91-c96e271d4d40-kube-api-access-69tqb\") pod \"watcher-db-create-8lkkh\" (UID: \"817d5ad4-17fa-4845-9a91-c96e271d4d40\") " pod="openstack/watcher-db-create-8lkkh" Oct 08 15:32:33 crc kubenswrapper[4945]: I1008 15:32:33.575003 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-8lkkh" Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.058228 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-8lkkh"] Oct 08 15:32:34 crc kubenswrapper[4945]: W1008 15:32:34.350496 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod817d5ad4_17fa_4845_9a91_c96e271d4d40.slice/crio-4bf9ba21557e6cffa91251b4ea09e0fb8ad357a38283ef52b14ed17fa716ab21 WatchSource:0}: Error finding container 4bf9ba21557e6cffa91251b4ea09e0fb8ad357a38283ef52b14ed17fa716ab21: Status 404 returned error can't find the container with id 4bf9ba21557e6cffa91251b4ea09e0fb8ad357a38283ef52b14ed17fa716ab21 Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.355523 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"420a6f35-1c79-416f-a7f4-8b9f46e337bc","Type":"ContainerStarted","Data":"c184f7ed9abbfb47a92b1823213f9ddd81745893d1429f145f4e55b0ae1c22de"} Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.355727 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.358658 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2481ce07-e2b2-4afb-b5f1-027bfefd0f38","Type":"ContainerStarted","Data":"4e408d857f77f1cce0c74bb1e177aeff00c716bca702c1487d3a63e4cdf3fba0"} Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.358855 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.360664 4945 generic.go:334] "Generic (PLEG): container finished" podID="4cc1f5af-e985-4b88-bb14-2d4cd9555188" containerID="db3fad6c030c3ba54c04dd337f0c8d6e47e0329111210fd589b34edee2e5ab0d" exitCode=0 Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.360732 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n5pnx" event={"ID":"4cc1f5af-e985-4b88-bb14-2d4cd9555188","Type":"ContainerDied","Data":"db3fad6c030c3ba54c04dd337f0c8d6e47e0329111210fd589b34edee2e5ab0d"} Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.363674 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec56f6d2-9c69-47f1-b834-169de890c512","Type":"ContainerStarted","Data":"8a07f234ae8ef9ef30b98aaf68ce0a8496f264fed9c6a782c9995e5b987b152b"} Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.364002 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.365874 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerStarted","Data":"ccf2d3f6875f8e248e85acba032360c8d1a918e0ac25f4a472e7c4df1f2cb75a"} Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.388332 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=56.500293422 podStartE2EDuration="59.388315014s" podCreationTimestamp="2025-10-08 15:31:35 +0000 UTC" firstStartedPulling="2025-10-08 15:31:56.212525136 +0000 UTC m=+1125.566440077" lastFinishedPulling="2025-10-08 15:31:59.100546768 +0000 UTC m=+1128.454461669" observedRunningTime="2025-10-08 15:32:34.385159005 +0000 UTC m=+1163.739073916" watchObservedRunningTime="2025-10-08 15:32:34.388315014 +0000 UTC m=+1163.742229915" Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.422324 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=56.124911867 podStartE2EDuration="58.422304769s" podCreationTimestamp="2025-10-08 15:31:36 +0000 UTC" firstStartedPulling="2025-10-08 15:31:56.805050925 +0000 UTC m=+1126.158965826" lastFinishedPulling="2025-10-08 15:31:59.102443827 +0000 UTC m=+1128.456358728" observedRunningTime="2025-10-08 15:32:34.414162814 +0000 UTC m=+1163.768077725" watchObservedRunningTime="2025-10-08 15:32:34.422304769 +0000 UTC m=+1163.776219680" Oct 08 15:32:34 crc kubenswrapper[4945]: I1008 15:32:34.450964 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=59.302330033 podStartE2EDuration="59.450949248s" podCreationTimestamp="2025-10-08 15:31:35 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.9529563 +0000 UTC m=+1128.306871201" lastFinishedPulling="2025-10-08 15:31:59.101575505 +0000 UTC m=+1128.455490416" observedRunningTime="2025-10-08 15:32:34.449521642 +0000 UTC m=+1163.803436553" watchObservedRunningTime="2025-10-08 15:32:34.450949248 +0000 UTC m=+1163.804864149" Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.378854 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-8lkkh" event={"ID":"817d5ad4-17fa-4845-9a91-c96e271d4d40","Type":"ContainerStarted","Data":"0ef9ba4567aa1ef7ddf1c20bafc2284544ee1b1bd89997c1b290f41a859c336e"} Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.379147 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-8lkkh" event={"ID":"817d5ad4-17fa-4845-9a91-c96e271d4d40","Type":"ContainerStarted","Data":"4bf9ba21557e6cffa91251b4ea09e0fb8ad357a38283ef52b14ed17fa716ab21"} Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.380972 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"c02ae4c023d21a9580526cd380f63839d44ade8525014c88dcfb46985113e8dc"} Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.407327 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-create-8lkkh" podStartSLOduration=2.407304511 podStartE2EDuration="2.407304511s" podCreationTimestamp="2025-10-08 15:32:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:35.398509701 +0000 UTC m=+1164.752424612" watchObservedRunningTime="2025-10-08 15:32:35.407304511 +0000 UTC m=+1164.761219432" Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.706915 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n5pnx" Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.806746 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86lqz\" (UniqueName: \"kubernetes.io/projected/4cc1f5af-e985-4b88-bb14-2d4cd9555188-kube-api-access-86lqz\") pod \"4cc1f5af-e985-4b88-bb14-2d4cd9555188\" (UID: \"4cc1f5af-e985-4b88-bb14-2d4cd9555188\") " Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.812444 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc1f5af-e985-4b88-bb14-2d4cd9555188-kube-api-access-86lqz" (OuterVolumeSpecName: "kube-api-access-86lqz") pod "4cc1f5af-e985-4b88-bb14-2d4cd9555188" (UID: "4cc1f5af-e985-4b88-bb14-2d4cd9555188"). InnerVolumeSpecName "kube-api-access-86lqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:35 crc kubenswrapper[4945]: I1008 15:32:35.908256 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86lqz\" (UniqueName: \"kubernetes.io/projected/4cc1f5af-e985-4b88-bb14-2d4cd9555188-kube-api-access-86lqz\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:36 crc kubenswrapper[4945]: I1008 15:32:36.393317 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"c1e0263f8bb1098f9a814492a798b1e505d2a424802cca1dd7f0165e5e7a7cd0"} Oct 08 15:32:36 crc kubenswrapper[4945]: I1008 15:32:36.394940 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n5pnx" Oct 08 15:32:36 crc kubenswrapper[4945]: I1008 15:32:36.394933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n5pnx" event={"ID":"4cc1f5af-e985-4b88-bb14-2d4cd9555188","Type":"ContainerDied","Data":"b1f812e0b2f91c9c4590008873d9e6b656bfa249a342ff3a9d6d46a74c4c2975"} Oct 08 15:32:36 crc kubenswrapper[4945]: I1008 15:32:36.395032 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1f812e0b2f91c9c4590008873d9e6b656bfa249a342ff3a9d6d46a74c4c2975" Oct 08 15:32:36 crc kubenswrapper[4945]: I1008 15:32:36.903540 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cwwpc" podUID="1b70322d-ae60-46b4-828b-1f4a01cae6b6" containerName="ovn-controller" probeResult="failure" output=< Oct 08 15:32:36 crc kubenswrapper[4945]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 15:32:36 crc kubenswrapper[4945]: > Oct 08 15:32:37 crc kubenswrapper[4945]: I1008 15:32:37.404207 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"80e954be63d95324c0b3e0851bddf82e58f069e3808107e66f3c2fb2112ec63a"} Oct 08 15:32:37 crc kubenswrapper[4945]: I1008 15:32:37.405794 4945 generic.go:334] "Generic (PLEG): container finished" podID="817d5ad4-17fa-4845-9a91-c96e271d4d40" containerID="0ef9ba4567aa1ef7ddf1c20bafc2284544ee1b1bd89997c1b290f41a859c336e" exitCode=0 Oct 08 15:32:37 crc kubenswrapper[4945]: I1008 15:32:37.405828 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-8lkkh" event={"ID":"817d5ad4-17fa-4845-9a91-c96e271d4d40","Type":"ContainerDied","Data":"0ef9ba4567aa1ef7ddf1c20bafc2284544ee1b1bd89997c1b290f41a859c336e"} Oct 08 15:32:38 crc kubenswrapper[4945]: I1008 15:32:38.419137 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"d06ef72fd5cdebe1a02c30be79d48e3dbd1e2ed2c34f81a175e97787789f9196"} Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.437593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-8lkkh" event={"ID":"817d5ad4-17fa-4845-9a91-c96e271d4d40","Type":"ContainerDied","Data":"4bf9ba21557e6cffa91251b4ea09e0fb8ad357a38283ef52b14ed17fa716ab21"} Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.438222 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bf9ba21557e6cffa91251b4ea09e0fb8ad357a38283ef52b14ed17fa716ab21" Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.495010 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-8lkkh" Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.582174 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69tqb\" (UniqueName: \"kubernetes.io/projected/817d5ad4-17fa-4845-9a91-c96e271d4d40-kube-api-access-69tqb\") pod \"817d5ad4-17fa-4845-9a91-c96e271d4d40\" (UID: \"817d5ad4-17fa-4845-9a91-c96e271d4d40\") " Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.587570 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/817d5ad4-17fa-4845-9a91-c96e271d4d40-kube-api-access-69tqb" (OuterVolumeSpecName: "kube-api-access-69tqb") pod "817d5ad4-17fa-4845-9a91-c96e271d4d40" (UID: "817d5ad4-17fa-4845-9a91-c96e271d4d40"). InnerVolumeSpecName "kube-api-access-69tqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.684437 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69tqb\" (UniqueName: \"kubernetes.io/projected/817d5ad4-17fa-4845-9a91-c96e271d4d40-kube-api-access-69tqb\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.999364 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-8z2zf"] Oct 08 15:32:40 crc kubenswrapper[4945]: E1008 15:32:40.999769 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc1f5af-e985-4b88-bb14-2d4cd9555188" containerName="mariadb-database-create" Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.999789 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc1f5af-e985-4b88-bb14-2d4cd9555188" containerName="mariadb-database-create" Oct 08 15:32:40 crc kubenswrapper[4945]: E1008 15:32:40.999814 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="817d5ad4-17fa-4845-9a91-c96e271d4d40" containerName="mariadb-database-create" Oct 08 15:32:40 crc kubenswrapper[4945]: I1008 15:32:40.999823 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="817d5ad4-17fa-4845-9a91-c96e271d4d40" containerName="mariadb-database-create" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.000043 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc1f5af-e985-4b88-bb14-2d4cd9555188" containerName="mariadb-database-create" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.000082 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="817d5ad4-17fa-4845-9a91-c96e271d4d40" containerName="mariadb-database-create" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.000758 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8z2zf" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.017743 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8z2zf"] Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.091351 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgb5q\" (UniqueName: \"kubernetes.io/projected/b9e17266-40ce-4c98-8f15-80e8da63774d-kube-api-access-zgb5q\") pod \"keystone-db-create-8z2zf\" (UID: \"b9e17266-40ce-4c98-8f15-80e8da63774d\") " pod="openstack/keystone-db-create-8z2zf" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.192545 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgb5q\" (UniqueName: \"kubernetes.io/projected/b9e17266-40ce-4c98-8f15-80e8da63774d-kube-api-access-zgb5q\") pod \"keystone-db-create-8z2zf\" (UID: \"b9e17266-40ce-4c98-8f15-80e8da63774d\") " pod="openstack/keystone-db-create-8z2zf" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.214545 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgb5q\" (UniqueName: \"kubernetes.io/projected/b9e17266-40ce-4c98-8f15-80e8da63774d-kube-api-access-zgb5q\") pod \"keystone-db-create-8z2zf\" (UID: \"b9e17266-40ce-4c98-8f15-80e8da63774d\") " pod="openstack/keystone-db-create-8z2zf" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.319046 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8z2zf" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.450927 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-8lkkh" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.793264 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8z2zf"] Oct 08 15:32:41 crc kubenswrapper[4945]: W1008 15:32:41.803301 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9e17266_40ce_4c98_8f15_80e8da63774d.slice/crio-e5c4fa1925424f314e7ac0db3b05f895788e48b7ea9a06294429e9f3875b7113 WatchSource:0}: Error finding container e5c4fa1925424f314e7ac0db3b05f895788e48b7ea9a06294429e9f3875b7113: Status 404 returned error can't find the container with id e5c4fa1925424f314e7ac0db3b05f895788e48b7ea9a06294429e9f3875b7113 Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.910426 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cwwpc" podUID="1b70322d-ae60-46b4-828b-1f4a01cae6b6" containerName="ovn-controller" probeResult="failure" output=< Oct 08 15:32:41 crc kubenswrapper[4945]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 15:32:41 crc kubenswrapper[4945]: > Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.935437 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:32:41 crc kubenswrapper[4945]: I1008 15:32:41.939738 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-67zmn" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.151671 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cwwpc-config-h8hhj"] Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.153470 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.157292 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.169383 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cwwpc-config-h8hhj"] Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.214358 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run-ovn\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.214642 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-scripts\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.214761 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-log-ovn\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.214799 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-additional-scripts\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.214877 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.214907 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvd2w\" (UniqueName: \"kubernetes.io/projected/a02041da-d509-43f6-95b6-b7ac679a88c7-kube-api-access-kvd2w\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.318444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-scripts\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.318736 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-log-ovn\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.318844 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-additional-scripts\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.318999 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.319181 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvd2w\" (UniqueName: \"kubernetes.io/projected/a02041da-d509-43f6-95b6-b7ac679a88c7-kube-api-access-kvd2w\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.319328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run-ovn\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.319828 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run-ovn\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.322864 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-scripts\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.323971 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-log-ovn\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.325042 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-additional-scripts\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.325288 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.345436 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvd2w\" (UniqueName: \"kubernetes.io/projected/a02041da-d509-43f6-95b6-b7ac679a88c7-kube-api-access-kvd2w\") pod \"ovn-controller-cwwpc-config-h8hhj\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.460365 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8z2zf" event={"ID":"b9e17266-40ce-4c98-8f15-80e8da63774d","Type":"ContainerStarted","Data":"c7fc8b011611ab6a3d06f1830a19fdb5489c5392921cac4a36df353c4271e3ba"} Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.461465 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8z2zf" event={"ID":"b9e17266-40ce-4c98-8f15-80e8da63774d","Type":"ContainerStarted","Data":"e5c4fa1925424f314e7ac0db3b05f895788e48b7ea9a06294429e9f3875b7113"} Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.475153 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:42 crc kubenswrapper[4945]: I1008 15:32:42.478654 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-8z2zf" podStartSLOduration=2.4786333640000002 podStartE2EDuration="2.478633364s" podCreationTimestamp="2025-10-08 15:32:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:42.470799327 +0000 UTC m=+1171.824714228" watchObservedRunningTime="2025-10-08 15:32:42.478633364 +0000 UTC m=+1171.832548265" Oct 08 15:32:43 crc kubenswrapper[4945]: I1008 15:32:43.469284 4945 generic.go:334] "Generic (PLEG): container finished" podID="b9e17266-40ce-4c98-8f15-80e8da63774d" containerID="c7fc8b011611ab6a3d06f1830a19fdb5489c5392921cac4a36df353c4271e3ba" exitCode=0 Oct 08 15:32:43 crc kubenswrapper[4945]: I1008 15:32:43.469373 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8z2zf" event={"ID":"b9e17266-40ce-4c98-8f15-80e8da63774d","Type":"ContainerDied","Data":"c7fc8b011611ab6a3d06f1830a19fdb5489c5392921cac4a36df353c4271e3ba"} Oct 08 15:32:43 crc kubenswrapper[4945]: I1008 15:32:43.471884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerStarted","Data":"874021f644a9ba63a2c58f6ab1031d5a95f522887f66d0d1fc1d83ef1e60787e"} Oct 08 15:32:43 crc kubenswrapper[4945]: I1008 15:32:43.607089 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cwwpc-config-h8hhj"] Oct 08 15:32:43 crc kubenswrapper[4945]: W1008 15:32:43.886214 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda02041da_d509_43f6_95b6_b7ac679a88c7.slice/crio-e009aafedfed5eec0f629f4d3cbe2809c1adcaeea4cc7b361b7cd4e35dadbff4 WatchSource:0}: Error finding container e009aafedfed5eec0f629f4d3cbe2809c1adcaeea4cc7b361b7cd4e35dadbff4: Status 404 returned error can't find the container with id e009aafedfed5eec0f629f4d3cbe2809c1adcaeea4cc7b361b7cd4e35dadbff4 Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.494018 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"68f4bb4451effad3744565c9e66344e284e598135f20c115ea20d5adcd8f4644"} Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.494424 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"0afaa33cf777ea91216c10768cac33970c06785ddc4a26f62a548fe0cb6b3f6e"} Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.498031 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cwwpc-config-h8hhj" event={"ID":"a02041da-d509-43f6-95b6-b7ac679a88c7","Type":"ContainerStarted","Data":"7cbbba4d77105c1ea7fdaa1f04b108f3689c6121e56befb96a2bc2ecb6f323cf"} Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.498078 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cwwpc-config-h8hhj" event={"ID":"a02041da-d509-43f6-95b6-b7ac679a88c7","Type":"ContainerStarted","Data":"e009aafedfed5eec0f629f4d3cbe2809c1adcaeea4cc7b361b7cd4e35dadbff4"} Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.518547 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.518609 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.521867 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.529324 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.2961041 podStartE2EDuration="1m1.529303167s" podCreationTimestamp="2025-10-08 15:31:43 +0000 UTC" firstStartedPulling="2025-10-08 15:31:58.926785869 +0000 UTC m=+1128.280700770" lastFinishedPulling="2025-10-08 15:32:43.159984936 +0000 UTC m=+1172.513899837" observedRunningTime="2025-10-08 15:32:44.524322592 +0000 UTC m=+1173.878237513" watchObservedRunningTime="2025-10-08 15:32:44.529303167 +0000 UTC m=+1173.883218068" Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.553068 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-cwwpc-config-h8hhj" podStartSLOduration=2.553044323 podStartE2EDuration="2.553044323s" podCreationTimestamp="2025-10-08 15:32:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:44.549839233 +0000 UTC m=+1173.903754134" watchObservedRunningTime="2025-10-08 15:32:44.553044323 +0000 UTC m=+1173.906959224" Oct 08 15:32:44 crc kubenswrapper[4945]: I1008 15:32:44.930185 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8z2zf" Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.067363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgb5q\" (UniqueName: \"kubernetes.io/projected/b9e17266-40ce-4c98-8f15-80e8da63774d-kube-api-access-zgb5q\") pod \"b9e17266-40ce-4c98-8f15-80e8da63774d\" (UID: \"b9e17266-40ce-4c98-8f15-80e8da63774d\") " Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.075548 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9e17266-40ce-4c98-8f15-80e8da63774d-kube-api-access-zgb5q" (OuterVolumeSpecName: "kube-api-access-zgb5q") pod "b9e17266-40ce-4c98-8f15-80e8da63774d" (UID: "b9e17266-40ce-4c98-8f15-80e8da63774d"). InnerVolumeSpecName "kube-api-access-zgb5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.168948 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgb5q\" (UniqueName: \"kubernetes.io/projected/b9e17266-40ce-4c98-8f15-80e8da63774d-kube-api-access-zgb5q\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.508692 4945 generic.go:334] "Generic (PLEG): container finished" podID="a02041da-d509-43f6-95b6-b7ac679a88c7" containerID="7cbbba4d77105c1ea7fdaa1f04b108f3689c6121e56befb96a2bc2ecb6f323cf" exitCode=0 Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.508803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cwwpc-config-h8hhj" event={"ID":"a02041da-d509-43f6-95b6-b7ac679a88c7","Type":"ContainerDied","Data":"7cbbba4d77105c1ea7fdaa1f04b108f3689c6121e56befb96a2bc2ecb6f323cf"} Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.513031 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8z2zf" Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.513037 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8z2zf" event={"ID":"b9e17266-40ce-4c98-8f15-80e8da63774d","Type":"ContainerDied","Data":"e5c4fa1925424f314e7ac0db3b05f895788e48b7ea9a06294429e9f3875b7113"} Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.513297 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5c4fa1925424f314e7ac0db3b05f895788e48b7ea9a06294429e9f3875b7113" Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.523825 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"65d00025a969dcfe3751c4189af32cd7ad58222852c69ceaf6a87f17a937435c"} Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.523867 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"85820f39cb2ee817ec373a809a3d115a01105c9c8bcc648b13cf8f8a43a36ba2"} Oct 08 15:32:45 crc kubenswrapper[4945]: I1008 15:32:45.525336 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:46 crc kubenswrapper[4945]: I1008 15:32:46.825690 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Oct 08 15:32:46 crc kubenswrapper[4945]: I1008 15:32:46.884350 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:46 crc kubenswrapper[4945]: I1008 15:32:46.957803 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-cwwpc" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.004857 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvd2w\" (UniqueName: \"kubernetes.io/projected/a02041da-d509-43f6-95b6-b7ac679a88c7-kube-api-access-kvd2w\") pod \"a02041da-d509-43f6-95b6-b7ac679a88c7\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.004914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-scripts\") pod \"a02041da-d509-43f6-95b6-b7ac679a88c7\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.005001 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run\") pod \"a02041da-d509-43f6-95b6-b7ac679a88c7\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.005090 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-additional-scripts\") pod \"a02041da-d509-43f6-95b6-b7ac679a88c7\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.005140 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-log-ovn\") pod \"a02041da-d509-43f6-95b6-b7ac679a88c7\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.005161 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run-ovn\") pod \"a02041da-d509-43f6-95b6-b7ac679a88c7\" (UID: \"a02041da-d509-43f6-95b6-b7ac679a88c7\") " Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.006881 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a02041da-d509-43f6-95b6-b7ac679a88c7" (UID: "a02041da-d509-43f6-95b6-b7ac679a88c7"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.007211 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a02041da-d509-43f6-95b6-b7ac679a88c7" (UID: "a02041da-d509-43f6-95b6-b7ac679a88c7"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.007284 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run" (OuterVolumeSpecName: "var-run") pod "a02041da-d509-43f6-95b6-b7ac679a88c7" (UID: "a02041da-d509-43f6-95b6-b7ac679a88c7"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.007665 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a02041da-d509-43f6-95b6-b7ac679a88c7" (UID: "a02041da-d509-43f6-95b6-b7ac679a88c7"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.008860 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-scripts" (OuterVolumeSpecName: "scripts") pod "a02041da-d509-43f6-95b6-b7ac679a88c7" (UID: "a02041da-d509-43f6-95b6-b7ac679a88c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.028330 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a02041da-d509-43f6-95b6-b7ac679a88c7-kube-api-access-kvd2w" (OuterVolumeSpecName: "kube-api-access-kvd2w") pod "a02041da-d509-43f6-95b6-b7ac679a88c7" (UID: "a02041da-d509-43f6-95b6-b7ac679a88c7"). InnerVolumeSpecName "kube-api-access-kvd2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.093397 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="420a6f35-1c79-416f-a7f4-8b9f46e337bc" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.107331 4945 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.107367 4945 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.107377 4945 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.107385 4945 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a02041da-d509-43f6-95b6-b7ac679a88c7-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.107395 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvd2w\" (UniqueName: \"kubernetes.io/projected/a02041da-d509-43f6-95b6-b7ac679a88c7-kube-api-access-kvd2w\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.107405 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a02041da-d509-43f6-95b6-b7ac679a88c7-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.524051 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.538850 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cwwpc-config-h8hhj" event={"ID":"a02041da-d509-43f6-95b6-b7ac679a88c7","Type":"ContainerDied","Data":"e009aafedfed5eec0f629f4d3cbe2809c1adcaeea4cc7b361b7cd4e35dadbff4"} Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.538886 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e009aafedfed5eec0f629f4d3cbe2809c1adcaeea4cc7b361b7cd4e35dadbff4" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.538904 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cwwpc-config-h8hhj" Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.542823 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"c9af6f86faa8a700db8efe7be5b19d17a67578274518240ec747ebc6586ff4ec"} Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.542856 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"66804ba76823fd0b25751781719198c9afc69b737b6d9d19ea8c5c0e3dc2a109"} Oct 08 15:32:47 crc kubenswrapper[4945]: I1008 15:32:47.870963 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:32:48 crc kubenswrapper[4945]: I1008 15:32:48.022577 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-cwwpc-config-h8hhj"] Oct 08 15:32:48 crc kubenswrapper[4945]: I1008 15:32:48.033903 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-cwwpc-config-h8hhj"] Oct 08 15:32:48 crc kubenswrapper[4945]: I1008 15:32:48.556105 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"33bbbcc935a88e67a3a90dd6e50920b1993520d69387373622aa801eb9d04c7f"} Oct 08 15:32:48 crc kubenswrapper[4945]: I1008 15:32:48.556351 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="prometheus" containerID="cri-o://e719bddc4017dc7bcdc4d982354494cbd4619fc87e3ca76e4cde77b0e9d7b204" gracePeriod=600 Oct 08 15:32:48 crc kubenswrapper[4945]: I1008 15:32:48.556813 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="thanos-sidecar" containerID="cri-o://874021f644a9ba63a2c58f6ab1031d5a95f522887f66d0d1fc1d83ef1e60787e" gracePeriod=600 Oct 08 15:32:48 crc kubenswrapper[4945]: I1008 15:32:48.556849 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="config-reloader" containerID="cri-o://ccf2d3f6875f8e248e85acba032360c8d1a918e0ac25f4a472e7c4df1f2cb75a" gracePeriod=600 Oct 08 15:32:49 crc kubenswrapper[4945]: I1008 15:32:49.184228 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:32:49 crc kubenswrapper[4945]: I1008 15:32:49.184447 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:32:49 crc kubenswrapper[4945]: I1008 15:32:49.518973 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.114:9090/-/ready\": dial tcp 10.217.0.114:9090: connect: connection refused" Oct 08 15:32:49 crc kubenswrapper[4945]: I1008 15:32:49.566123 4945 generic.go:334] "Generic (PLEG): container finished" podID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerID="874021f644a9ba63a2c58f6ab1031d5a95f522887f66d0d1fc1d83ef1e60787e" exitCode=0 Oct 08 15:32:49 crc kubenswrapper[4945]: I1008 15:32:49.566158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerDied","Data":"874021f644a9ba63a2c58f6ab1031d5a95f522887f66d0d1fc1d83ef1e60787e"} Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.034308 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a02041da-d509-43f6-95b6-b7ac679a88c7" path="/var/lib/kubelet/pods/a02041da-d509-43f6-95b6-b7ac679a88c7/volumes" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.576176 4945 generic.go:334] "Generic (PLEG): container finished" podID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerID="ccf2d3f6875f8e248e85acba032360c8d1a918e0ac25f4a472e7c4df1f2cb75a" exitCode=0 Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.576503 4945 generic.go:334] "Generic (PLEG): container finished" podID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerID="e719bddc4017dc7bcdc4d982354494cbd4619fc87e3ca76e4cde77b0e9d7b204" exitCode=0 Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.576324 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerDied","Data":"ccf2d3f6875f8e248e85acba032360c8d1a918e0ac25f4a472e7c4df1f2cb75a"} Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.576580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerDied","Data":"e719bddc4017dc7bcdc4d982354494cbd4619fc87e3ca76e4cde77b0e9d7b204"} Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.588607 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"8e667f8028a488db8daf91e9711c37c26f6e938365819c014fcddb786832d70b"} Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.766019 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.881423 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config-out\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.881549 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.881600 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-tls-assets\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.881635 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q9k6\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-kube-api-access-8q9k6\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.881857 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.881958 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-web-config\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.882135 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-prometheus-metric-storage-rulefiles-0\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.882184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-thanos-prometheus-http-client-file\") pod \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\" (UID: \"1b942286-9a63-4a9f-9aa9-822ad8f3a34a\") " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.889615 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config-out" (OuterVolumeSpecName: "config-out") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.891505 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.891747 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.895681 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-kube-api-access-8q9k6" (OuterVolumeSpecName: "kube-api-access-8q9k6") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "kube-api-access-8q9k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.897490 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.901543 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config" (OuterVolumeSpecName: "config") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.915821 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.926376 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-web-config" (OuterVolumeSpecName: "web-config") pod "1b942286-9a63-4a9f-9aa9-822ad8f3a34a" (UID: "1b942286-9a63-4a9f-9aa9-822ad8f3a34a"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985051 4945 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985080 4945 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985094 4945 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config-out\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985121 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985136 4945 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985146 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q9k6\" (UniqueName: \"kubernetes.io/projected/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-kube-api-access-8q9k6\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985186 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") on node \"crc\" " Oct 08 15:32:50 crc kubenswrapper[4945]: I1008 15:32:50.985199 4945 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1b942286-9a63-4a9f-9aa9-822ad8f3a34a-web-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.022201 4945 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.022376 4945 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e") on node "crc" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.086649 4945 reconciler_common.go:293] "Volume detached for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.089102 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f58f-account-create-t2jxv"] Oct 08 15:32:51 crc kubenswrapper[4945]: E1008 15:32:51.089519 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="init-config-reloader" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.089541 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="init-config-reloader" Oct 08 15:32:51 crc kubenswrapper[4945]: E1008 15:32:51.089568 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="thanos-sidecar" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.089576 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="thanos-sidecar" Oct 08 15:32:51 crc kubenswrapper[4945]: E1008 15:32:51.089592 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="prometheus" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.089601 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="prometheus" Oct 08 15:32:51 crc kubenswrapper[4945]: E1008 15:32:51.089615 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e17266-40ce-4c98-8f15-80e8da63774d" containerName="mariadb-database-create" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.089622 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e17266-40ce-4c98-8f15-80e8da63774d" containerName="mariadb-database-create" Oct 08 15:32:51 crc kubenswrapper[4945]: E1008 15:32:51.089650 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="config-reloader" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.089658 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="config-reloader" Oct 08 15:32:51 crc kubenswrapper[4945]: E1008 15:32:51.089677 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a02041da-d509-43f6-95b6-b7ac679a88c7" containerName="ovn-config" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.089685 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a02041da-d509-43f6-95b6-b7ac679a88c7" containerName="ovn-config" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.095361 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="prometheus" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.095420 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="thanos-sidecar" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.095443 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a02041da-d509-43f6-95b6-b7ac679a88c7" containerName="ovn-config" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.095459 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9e17266-40ce-4c98-8f15-80e8da63774d" containerName="mariadb-database-create" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.095475 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" containerName="config-reloader" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.096192 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f58f-account-create-t2jxv" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.097451 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f58f-account-create-t2jxv"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.098338 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.187917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zvc9\" (UniqueName: \"kubernetes.io/projected/62fa8723-fb82-45bf-9750-2a167cdec803-kube-api-access-5zvc9\") pod \"keystone-f58f-account-create-t2jxv\" (UID: \"62fa8723-fb82-45bf-9750-2a167cdec803\") " pod="openstack/keystone-f58f-account-create-t2jxv" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.289564 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zvc9\" (UniqueName: \"kubernetes.io/projected/62fa8723-fb82-45bf-9750-2a167cdec803-kube-api-access-5zvc9\") pod \"keystone-f58f-account-create-t2jxv\" (UID: \"62fa8723-fb82-45bf-9750-2a167cdec803\") " pod="openstack/keystone-f58f-account-create-t2jxv" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.313366 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zvc9\" (UniqueName: \"kubernetes.io/projected/62fa8723-fb82-45bf-9750-2a167cdec803-kube-api-access-5zvc9\") pod \"keystone-f58f-account-create-t2jxv\" (UID: \"62fa8723-fb82-45bf-9750-2a167cdec803\") " pod="openstack/keystone-f58f-account-create-t2jxv" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.384442 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-e917-account-create-5nmhp"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.385754 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e917-account-create-5nmhp" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.394197 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e917-account-create-5nmhp"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.397132 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.414778 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f58f-account-create-t2jxv" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.493580 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqkjb\" (UniqueName: \"kubernetes.io/projected/b99dc5f0-37b6-45ed-8a7c-8b2299405792-kube-api-access-sqkjb\") pod \"placement-e917-account-create-5nmhp\" (UID: \"b99dc5f0-37b6-45ed-8a7c-8b2299405792\") " pod="openstack/placement-e917-account-create-5nmhp" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.595761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqkjb\" (UniqueName: \"kubernetes.io/projected/b99dc5f0-37b6-45ed-8a7c-8b2299405792-kube-api-access-sqkjb\") pod \"placement-e917-account-create-5nmhp\" (UID: \"b99dc5f0-37b6-45ed-8a7c-8b2299405792\") " pod="openstack/placement-e917-account-create-5nmhp" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.602523 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1b942286-9a63-4a9f-9aa9-822ad8f3a34a","Type":"ContainerDied","Data":"3d1f8df3374c53cf40964f5e8a5d9cdac1321d84a9d70d13809f620a609f3544"} Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.602603 4945 scope.go:117] "RemoveContainer" containerID="874021f644a9ba63a2c58f6ab1031d5a95f522887f66d0d1fc1d83ef1e60787e" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.602541 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.615559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqkjb\" (UniqueName: \"kubernetes.io/projected/b99dc5f0-37b6-45ed-8a7c-8b2299405792-kube-api-access-sqkjb\") pod \"placement-e917-account-create-5nmhp\" (UID: \"b99dc5f0-37b6-45ed-8a7c-8b2299405792\") " pod="openstack/placement-e917-account-create-5nmhp" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.619670 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"9c074028c2a2debc1457ab0214f1dd27f67d8cb24e7557956127837fbe62cd32"} Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.619708 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"cf2f4d428f92b62e2f8fee6b326df2952efb2027086dbd9b91e0b2464cdd8635"} Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.619718 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e8011e91-9053-4a34-a58a-7d046d51536f","Type":"ContainerStarted","Data":"2568cc951872063824ce3b028714e6b9b6e36f56d63d6b25740e7e08e8bcb2ef"} Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.621497 4945 scope.go:117] "RemoveContainer" containerID="ccf2d3f6875f8e248e85acba032360c8d1a918e0ac25f4a472e7c4df1f2cb75a" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.665648 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=22.892716231 podStartE2EDuration="38.665631422s" podCreationTimestamp="2025-10-08 15:32:13 +0000 UTC" firstStartedPulling="2025-10-08 15:32:31.019674492 +0000 UTC m=+1160.373589403" lastFinishedPulling="2025-10-08 15:32:46.792589693 +0000 UTC m=+1176.146504594" observedRunningTime="2025-10-08 15:32:51.655004466 +0000 UTC m=+1181.008919367" watchObservedRunningTime="2025-10-08 15:32:51.665631422 +0000 UTC m=+1181.019546323" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.709268 4945 scope.go:117] "RemoveContainer" containerID="e719bddc4017dc7bcdc4d982354494cbd4619fc87e3ca76e4cde77b0e9d7b204" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.738148 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.739853 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e917-account-create-5nmhp" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.758405 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.777445 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.783172 4945 scope.go:117] "RemoveContainer" containerID="a07652c59d10c5c9ff261dfa0a5d05794c80ce3a34b1b4a4f5d2001b7fa1ab53" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.792177 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.792303 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.797027 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.797414 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.798133 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.798289 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.799016 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.799293 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-qmfdz" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.804825 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.868302 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f58f-account-create-t2jxv"] Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903083 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903292 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903334 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903396 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903453 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903536 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903564 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4fwf\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-kube-api-access-q4fwf\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:51 crc kubenswrapper[4945]: I1008 15:32:51.903606 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.004923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005210 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4fwf\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-kube-api-access-q4fwf\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005272 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005295 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005326 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005356 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005384 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005408 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.005477 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.014956 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.016226 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.025762 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.028291 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.031440 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.031846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.038210 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.042054 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.043745 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.051845 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4fwf\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-kube-api-access-q4fwf\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.061527 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b942286-9a63-4a9f-9aa9-822ad8f3a34a" path="/var/lib/kubelet/pods/1b942286-9a63-4a9f-9aa9-822ad8f3a34a/volumes" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.067036 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.067074 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9dcd0c273e01834b6d6b9a81fac4d1c0f2ca078125a1dee333b69bb46fb7d5e3/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.096174 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d9c464d5c-8j67t"] Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.097600 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.100865 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.109043 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.109128 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-svc\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.109191 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-config\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.109352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-swift-storage-0\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.109458 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.109524 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fncz6\" (UniqueName: \"kubernetes.io/projected/ade81fed-c2cb-46a1-9623-96bcc42b65b7-kube-api-access-fncz6\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.113148 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d9c464d5c-8j67t"] Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.209902 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e917-account-create-5nmhp"] Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.214047 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.214131 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-svc\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.214219 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-config\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.214281 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-swift-storage-0\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.214314 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.214360 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fncz6\" (UniqueName: \"kubernetes.io/projected/ade81fed-c2cb-46a1-9623-96bcc42b65b7-kube-api-access-fncz6\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.215133 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-svc\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.215598 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-swift-storage-0\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.215788 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.219664 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-config\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.228939 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.241235 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fncz6\" (UniqueName: \"kubernetes.io/projected/ade81fed-c2cb-46a1-9623-96bcc42b65b7-kube-api-access-fncz6\") pod \"dnsmasq-dns-6d9c464d5c-8j67t\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.332563 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.429641 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.434818 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.630958 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f58f-account-create-t2jxv" event={"ID":"62fa8723-fb82-45bf-9750-2a167cdec803","Type":"ContainerStarted","Data":"3bb8cd25df2313b3f4ae8d27722fd83a2e72e6de26a37b51a3900ca4dda4111b"} Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.631224 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f58f-account-create-t2jxv" event={"ID":"62fa8723-fb82-45bf-9750-2a167cdec803","Type":"ContainerStarted","Data":"082f6f770a0c22171ff067b51513a5d05fe809f7a9d75b7702f624e8929a53b3"} Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.634576 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e917-account-create-5nmhp" event={"ID":"b99dc5f0-37b6-45ed-8a7c-8b2299405792","Type":"ContainerStarted","Data":"3f0516edf2b32c59f0674abb53b6e769f259fcaf0649fd32cdaca1e406c2075e"} Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.634597 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e917-account-create-5nmhp" event={"ID":"b99dc5f0-37b6-45ed-8a7c-8b2299405792","Type":"ContainerStarted","Data":"77c4124da31356c6813a109485116d0f9b3ef1f5ccda4cebb206b5c0e2067a2f"} Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.654210 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-f58f-account-create-t2jxv" podStartSLOduration=1.6541929149999999 podStartE2EDuration="1.654192915s" podCreationTimestamp="2025-10-08 15:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:52.652683777 +0000 UTC m=+1182.006598678" watchObservedRunningTime="2025-10-08 15:32:52.654192915 +0000 UTC m=+1182.008107806" Oct 08 15:32:52 crc kubenswrapper[4945]: I1008 15:32:52.933559 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d9c464d5c-8j67t"] Oct 08 15:32:52 crc kubenswrapper[4945]: W1008 15:32:52.939904 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podade81fed_c2cb_46a1_9623_96bcc42b65b7.slice/crio-3060ab35805dc8b9be051b8d73ae26ec36e802cca3b4fbdbeaa84d5b6c39542e WatchSource:0}: Error finding container 3060ab35805dc8b9be051b8d73ae26ec36e802cca3b4fbdbeaa84d5b6c39542e: Status 404 returned error can't find the container with id 3060ab35805dc8b9be051b8d73ae26ec36e802cca3b4fbdbeaa84d5b6c39542e Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.013577 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 15:32:53 crc kubenswrapper[4945]: W1008 15:32:53.033579 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode540f1d4_87f5_4ddf_a9a0_d8141141b9d5.slice/crio-e02aa4b57adaaf2ba323f3d8cd27a5c3b0789317399f84f197fe9cac0c4ca889 WatchSource:0}: Error finding container e02aa4b57adaaf2ba323f3d8cd27a5c3b0789317399f84f197fe9cac0c4ca889: Status 404 returned error can't find the container with id e02aa4b57adaaf2ba323f3d8cd27a5c3b0789317399f84f197fe9cac0c4ca889 Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.270124 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-8ae8-account-create-655qj"] Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.271381 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-8ae8-account-create-655qj" Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.282183 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.285867 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-8ae8-account-create-655qj"] Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.337944 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjrlh\" (UniqueName: \"kubernetes.io/projected/b2403ece-f576-45c1-b7cd-0547eb470c75-kube-api-access-xjrlh\") pod \"watcher-8ae8-account-create-655qj\" (UID: \"b2403ece-f576-45c1-b7cd-0547eb470c75\") " pod="openstack/watcher-8ae8-account-create-655qj" Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.439765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjrlh\" (UniqueName: \"kubernetes.io/projected/b2403ece-f576-45c1-b7cd-0547eb470c75-kube-api-access-xjrlh\") pod \"watcher-8ae8-account-create-655qj\" (UID: \"b2403ece-f576-45c1-b7cd-0547eb470c75\") " pod="openstack/watcher-8ae8-account-create-655qj" Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.455918 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjrlh\" (UniqueName: \"kubernetes.io/projected/b2403ece-f576-45c1-b7cd-0547eb470c75-kube-api-access-xjrlh\") pod \"watcher-8ae8-account-create-655qj\" (UID: \"b2403ece-f576-45c1-b7cd-0547eb470c75\") " pod="openstack/watcher-8ae8-account-create-655qj" Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.600311 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-8ae8-account-create-655qj" Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.642710 4945 generic.go:334] "Generic (PLEG): container finished" podID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerID="811e0a6786dd761db08ddf61a496caaf55bd7adfd28f033b5683c051766e9e55" exitCode=0 Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.643361 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" event={"ID":"ade81fed-c2cb-46a1-9623-96bcc42b65b7","Type":"ContainerDied","Data":"811e0a6786dd761db08ddf61a496caaf55bd7adfd28f033b5683c051766e9e55"} Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.643395 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" event={"ID":"ade81fed-c2cb-46a1-9623-96bcc42b65b7","Type":"ContainerStarted","Data":"3060ab35805dc8b9be051b8d73ae26ec36e802cca3b4fbdbeaa84d5b6c39542e"} Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.684615 4945 generic.go:334] "Generic (PLEG): container finished" podID="62fa8723-fb82-45bf-9750-2a167cdec803" containerID="3bb8cd25df2313b3f4ae8d27722fd83a2e72e6de26a37b51a3900ca4dda4111b" exitCode=0 Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.684715 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f58f-account-create-t2jxv" event={"ID":"62fa8723-fb82-45bf-9750-2a167cdec803","Type":"ContainerDied","Data":"3bb8cd25df2313b3f4ae8d27722fd83a2e72e6de26a37b51a3900ca4dda4111b"} Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.688869 4945 generic.go:334] "Generic (PLEG): container finished" podID="b99dc5f0-37b6-45ed-8a7c-8b2299405792" containerID="3f0516edf2b32c59f0674abb53b6e769f259fcaf0649fd32cdaca1e406c2075e" exitCode=0 Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.688946 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e917-account-create-5nmhp" event={"ID":"b99dc5f0-37b6-45ed-8a7c-8b2299405792","Type":"ContainerDied","Data":"3f0516edf2b32c59f0674abb53b6e769f259fcaf0649fd32cdaca1e406c2075e"} Oct 08 15:32:53 crc kubenswrapper[4945]: I1008 15:32:53.694818 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerStarted","Data":"e02aa4b57adaaf2ba323f3d8cd27a5c3b0789317399f84f197fe9cac0c4ca889"} Oct 08 15:32:54 crc kubenswrapper[4945]: I1008 15:32:54.098624 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-8ae8-account-create-655qj"] Oct 08 15:32:54 crc kubenswrapper[4945]: I1008 15:32:54.704985 4945 generic.go:334] "Generic (PLEG): container finished" podID="b2403ece-f576-45c1-b7cd-0547eb470c75" containerID="011272c4b0c1eac555dfdae3036d6f495a71bbf710690bcd474820004ddb89b9" exitCode=0 Oct 08 15:32:54 crc kubenswrapper[4945]: I1008 15:32:54.705039 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-8ae8-account-create-655qj" event={"ID":"b2403ece-f576-45c1-b7cd-0547eb470c75","Type":"ContainerDied","Data":"011272c4b0c1eac555dfdae3036d6f495a71bbf710690bcd474820004ddb89b9"} Oct 08 15:32:54 crc kubenswrapper[4945]: I1008 15:32:54.705062 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-8ae8-account-create-655qj" event={"ID":"b2403ece-f576-45c1-b7cd-0547eb470c75","Type":"ContainerStarted","Data":"24b16b1fee268e443c889b80087dbbf06cc1ee13a5c882ba380227b346600ba5"} Oct 08 15:32:54 crc kubenswrapper[4945]: I1008 15:32:54.707723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" event={"ID":"ade81fed-c2cb-46a1-9623-96bcc42b65b7","Type":"ContainerStarted","Data":"31f9c9a3ea6dcbccb34ddd8909963c1f4777115961eb032e19e6244e53314035"} Oct 08 15:32:54 crc kubenswrapper[4945]: I1008 15:32:54.708449 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:32:54 crc kubenswrapper[4945]: I1008 15:32:54.769233 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" podStartSLOduration=2.769211415 podStartE2EDuration="2.769211415s" podCreationTimestamp="2025-10-08 15:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:32:54.762421635 +0000 UTC m=+1184.116336556" watchObservedRunningTime="2025-10-08 15:32:54.769211415 +0000 UTC m=+1184.123126326" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.421735 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e917-account-create-5nmhp" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.429345 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f58f-account-create-t2jxv" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.577121 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zvc9\" (UniqueName: \"kubernetes.io/projected/62fa8723-fb82-45bf-9750-2a167cdec803-kube-api-access-5zvc9\") pod \"62fa8723-fb82-45bf-9750-2a167cdec803\" (UID: \"62fa8723-fb82-45bf-9750-2a167cdec803\") " Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.577274 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqkjb\" (UniqueName: \"kubernetes.io/projected/b99dc5f0-37b6-45ed-8a7c-8b2299405792-kube-api-access-sqkjb\") pod \"b99dc5f0-37b6-45ed-8a7c-8b2299405792\" (UID: \"b99dc5f0-37b6-45ed-8a7c-8b2299405792\") " Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.583301 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fa8723-fb82-45bf-9750-2a167cdec803-kube-api-access-5zvc9" (OuterVolumeSpecName: "kube-api-access-5zvc9") pod "62fa8723-fb82-45bf-9750-2a167cdec803" (UID: "62fa8723-fb82-45bf-9750-2a167cdec803"). InnerVolumeSpecName "kube-api-access-5zvc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.587352 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b99dc5f0-37b6-45ed-8a7c-8b2299405792-kube-api-access-sqkjb" (OuterVolumeSpecName: "kube-api-access-sqkjb") pod "b99dc5f0-37b6-45ed-8a7c-8b2299405792" (UID: "b99dc5f0-37b6-45ed-8a7c-8b2299405792"). InnerVolumeSpecName "kube-api-access-sqkjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.679315 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqkjb\" (UniqueName: \"kubernetes.io/projected/b99dc5f0-37b6-45ed-8a7c-8b2299405792-kube-api-access-sqkjb\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.679574 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zvc9\" (UniqueName: \"kubernetes.io/projected/62fa8723-fb82-45bf-9750-2a167cdec803-kube-api-access-5zvc9\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.717902 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f58f-account-create-t2jxv" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.717884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f58f-account-create-t2jxv" event={"ID":"62fa8723-fb82-45bf-9750-2a167cdec803","Type":"ContainerDied","Data":"082f6f770a0c22171ff067b51513a5d05fe809f7a9d75b7702f624e8929a53b3"} Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.718076 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="082f6f770a0c22171ff067b51513a5d05fe809f7a9d75b7702f624e8929a53b3" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.719316 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e917-account-create-5nmhp" event={"ID":"b99dc5f0-37b6-45ed-8a7c-8b2299405792","Type":"ContainerDied","Data":"77c4124da31356c6813a109485116d0f9b3ef1f5ccda4cebb206b5c0e2067a2f"} Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.719346 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77c4124da31356c6813a109485116d0f9b3ef1f5ccda4cebb206b5c0e2067a2f" Oct 08 15:32:55 crc kubenswrapper[4945]: I1008 15:32:55.719408 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e917-account-create-5nmhp" Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.094401 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-8ae8-account-create-655qj" Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.186442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjrlh\" (UniqueName: \"kubernetes.io/projected/b2403ece-f576-45c1-b7cd-0547eb470c75-kube-api-access-xjrlh\") pod \"b2403ece-f576-45c1-b7cd-0547eb470c75\" (UID: \"b2403ece-f576-45c1-b7cd-0547eb470c75\") " Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.192458 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2403ece-f576-45c1-b7cd-0547eb470c75-kube-api-access-xjrlh" (OuterVolumeSpecName: "kube-api-access-xjrlh") pod "b2403ece-f576-45c1-b7cd-0547eb470c75" (UID: "b2403ece-f576-45c1-b7cd-0547eb470c75"). InnerVolumeSpecName "kube-api-access-xjrlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.288816 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjrlh\" (UniqueName: \"kubernetes.io/projected/b2403ece-f576-45c1-b7cd-0547eb470c75-kube-api-access-xjrlh\") on node \"crc\" DevicePath \"\"" Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.728706 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerStarted","Data":"81c5bd76ee389f38023724b07d0a44bc55ee8f44508d56fe54a7e0e51d2e56f8"} Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.730379 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-8ae8-account-create-655qj" event={"ID":"b2403ece-f576-45c1-b7cd-0547eb470c75","Type":"ContainerDied","Data":"24b16b1fee268e443c889b80087dbbf06cc1ee13a5c882ba380227b346600ba5"} Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.730418 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24b16b1fee268e443c889b80087dbbf06cc1ee13a5c882ba380227b346600ba5" Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.730470 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-8ae8-account-create-655qj" Oct 08 15:32:56 crc kubenswrapper[4945]: I1008 15:32:56.823306 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Oct 08 15:32:57 crc kubenswrapper[4945]: I1008 15:32:57.090857 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="420a6f35-1c79-416f-a7f4-8b9f46e337bc" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 08 15:32:57 crc kubenswrapper[4945]: I1008 15:32:57.523094 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Oct 08 15:33:02 crc kubenswrapper[4945]: I1008 15:33:02.431934 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:33:02 crc kubenswrapper[4945]: I1008 15:33:02.502390 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b89b7f545-vmj78"] Oct 08 15:33:02 crc kubenswrapper[4945]: I1008 15:33:02.502634 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" podUID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerName="dnsmasq-dns" containerID="cri-o://b7cd398e2b71fc245f28a3235d309f80fb40f4f5747176ebdaf555f8bc1c2cf5" gracePeriod=10 Oct 08 15:33:02 crc kubenswrapper[4945]: I1008 15:33:02.791273 4945 generic.go:334] "Generic (PLEG): container finished" podID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerID="b7cd398e2b71fc245f28a3235d309f80fb40f4f5747176ebdaf555f8bc1c2cf5" exitCode=0 Oct 08 15:33:02 crc kubenswrapper[4945]: I1008 15:33:02.791401 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" event={"ID":"acf8ac52-ea0c-4ea9-9a19-616edcd554eb","Type":"ContainerDied","Data":"b7cd398e2b71fc245f28a3235d309f80fb40f4f5747176ebdaf555f8bc1c2cf5"} Oct 08 15:33:02 crc kubenswrapper[4945]: I1008 15:33:02.793903 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerDied","Data":"81c5bd76ee389f38023724b07d0a44bc55ee8f44508d56fe54a7e0e51d2e56f8"} Oct 08 15:33:02 crc kubenswrapper[4945]: I1008 15:33:02.793907 4945 generic.go:334] "Generic (PLEG): container finished" podID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerID="81c5bd76ee389f38023724b07d0a44bc55ee8f44508d56fe54a7e0e51d2e56f8" exitCode=0 Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.022100 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.107310 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-dns-svc\") pod \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.107372 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sqsv\" (UniqueName: \"kubernetes.io/projected/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-kube-api-access-2sqsv\") pod \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.107415 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-config\") pod \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.107448 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-nb\") pod \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.107469 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-sb\") pod \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\" (UID: \"acf8ac52-ea0c-4ea9-9a19-616edcd554eb\") " Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.110445 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-kube-api-access-2sqsv" (OuterVolumeSpecName: "kube-api-access-2sqsv") pod "acf8ac52-ea0c-4ea9-9a19-616edcd554eb" (UID: "acf8ac52-ea0c-4ea9-9a19-616edcd554eb"). InnerVolumeSpecName "kube-api-access-2sqsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.150846 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-config" (OuterVolumeSpecName: "config") pod "acf8ac52-ea0c-4ea9-9a19-616edcd554eb" (UID: "acf8ac52-ea0c-4ea9-9a19-616edcd554eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.155666 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "acf8ac52-ea0c-4ea9-9a19-616edcd554eb" (UID: "acf8ac52-ea0c-4ea9-9a19-616edcd554eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.155713 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "acf8ac52-ea0c-4ea9-9a19-616edcd554eb" (UID: "acf8ac52-ea0c-4ea9-9a19-616edcd554eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.156149 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "acf8ac52-ea0c-4ea9-9a19-616edcd554eb" (UID: "acf8ac52-ea0c-4ea9-9a19-616edcd554eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.209737 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.209776 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.209789 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.210139 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sqsv\" (UniqueName: \"kubernetes.io/projected/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-kube-api-access-2sqsv\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.210156 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf8ac52-ea0c-4ea9-9a19-616edcd554eb-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.805663 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.805690 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b89b7f545-vmj78" event={"ID":"acf8ac52-ea0c-4ea9-9a19-616edcd554eb","Type":"ContainerDied","Data":"2de36999dfb27382d489c426a443d5806a5e1ffc99eeae999c8b8ac3f1170bd7"} Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.806173 4945 scope.go:117] "RemoveContainer" containerID="b7cd398e2b71fc245f28a3235d309f80fb40f4f5747176ebdaf555f8bc1c2cf5" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.807747 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerStarted","Data":"e93d65b01947105be03f6d85d3cb05be4b1a2bd6929ee804b1afffa14db7fd1f"} Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.837073 4945 scope.go:117] "RemoveContainer" containerID="9ac990b826437d87d72086f7681687339824a7e9e25cb806c64d7af7353331c9" Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.857916 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b89b7f545-vmj78"] Oct 08 15:33:03 crc kubenswrapper[4945]: I1008 15:33:03.866496 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b89b7f545-vmj78"] Oct 08 15:33:04 crc kubenswrapper[4945]: I1008 15:33:04.039793 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" path="/var/lib/kubelet/pods/acf8ac52-ea0c-4ea9-9a19-616edcd554eb/volumes" Oct 08 15:33:06 crc kubenswrapper[4945]: I1008 15:33:06.828472 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 15:33:06 crc kubenswrapper[4945]: I1008 15:33:06.837560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerStarted","Data":"f1f98955cf4183a5a254adcd49d46fb488561128e63fdb23c2e396868cbf0d1c"} Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.092399 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.249149 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-gzvqb"] Oct 08 15:33:07 crc kubenswrapper[4945]: E1008 15:33:07.249756 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerName="init" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.249776 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerName="init" Oct 08 15:33:07 crc kubenswrapper[4945]: E1008 15:33:07.249793 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fa8723-fb82-45bf-9750-2a167cdec803" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.249801 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fa8723-fb82-45bf-9750-2a167cdec803" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: E1008 15:33:07.249817 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2403ece-f576-45c1-b7cd-0547eb470c75" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.249825 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2403ece-f576-45c1-b7cd-0547eb470c75" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: E1008 15:33:07.249839 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerName="dnsmasq-dns" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.249846 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerName="dnsmasq-dns" Oct 08 15:33:07 crc kubenswrapper[4945]: E1008 15:33:07.249876 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99dc5f0-37b6-45ed-8a7c-8b2299405792" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.249886 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99dc5f0-37b6-45ed-8a7c-8b2299405792" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.250090 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99dc5f0-37b6-45ed-8a7c-8b2299405792" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.250128 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fa8723-fb82-45bf-9750-2a167cdec803" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.250141 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2403ece-f576-45c1-b7cd-0547eb470c75" containerName="mariadb-account-create" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.250157 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="acf8ac52-ea0c-4ea9-9a19-616edcd554eb" containerName="dnsmasq-dns" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.250796 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gzvqb" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.257294 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-gzvqb"] Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.356736 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-gp4z9"] Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.358683 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gp4z9" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.368068 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gp4z9"] Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.390390 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gmcx\" (UniqueName: \"kubernetes.io/projected/549f6009-8c9d-4fbb-890d-236b5c93c3fa-kube-api-access-9gmcx\") pod \"cinder-db-create-gzvqb\" (UID: \"549f6009-8c9d-4fbb-890d-236b5c93c3fa\") " pod="openstack/cinder-db-create-gzvqb" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.411722 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-978m6"] Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.413096 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.415198 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.415453 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.415697 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.415901 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7rf" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.435384 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-978m6"] Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.491760 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-config-data\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.491853 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7lzm\" (UniqueName: \"kubernetes.io/projected/6d7796c3-c17f-4795-b874-d5196d0deab2-kube-api-access-z7lzm\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.491917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-combined-ca-bundle\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.491940 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gmcx\" (UniqueName: \"kubernetes.io/projected/549f6009-8c9d-4fbb-890d-236b5c93c3fa-kube-api-access-9gmcx\") pod \"cinder-db-create-gzvqb\" (UID: \"549f6009-8c9d-4fbb-890d-236b5c93c3fa\") " pod="openstack/cinder-db-create-gzvqb" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.491987 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fld2t\" (UniqueName: \"kubernetes.io/projected/2c95dc54-6062-4f5f-8ffa-79cd1a23749b-kube-api-access-fld2t\") pod \"barbican-db-create-gp4z9\" (UID: \"2c95dc54-6062-4f5f-8ffa-79cd1a23749b\") " pod="openstack/barbican-db-create-gp4z9" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.510698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gmcx\" (UniqueName: \"kubernetes.io/projected/549f6009-8c9d-4fbb-890d-236b5c93c3fa-kube-api-access-9gmcx\") pod \"cinder-db-create-gzvqb\" (UID: \"549f6009-8c9d-4fbb-890d-236b5c93c3fa\") " pod="openstack/cinder-db-create-gzvqb" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.523326 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.593980 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-combined-ca-bundle\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.594087 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fld2t\" (UniqueName: \"kubernetes.io/projected/2c95dc54-6062-4f5f-8ffa-79cd1a23749b-kube-api-access-fld2t\") pod \"barbican-db-create-gp4z9\" (UID: \"2c95dc54-6062-4f5f-8ffa-79cd1a23749b\") " pod="openstack/barbican-db-create-gp4z9" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.594295 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-config-data\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.594393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7lzm\" (UniqueName: \"kubernetes.io/projected/6d7796c3-c17f-4795-b874-d5196d0deab2-kube-api-access-z7lzm\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.599582 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-config-data\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.599950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-combined-ca-bundle\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.612372 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fld2t\" (UniqueName: \"kubernetes.io/projected/2c95dc54-6062-4f5f-8ffa-79cd1a23749b-kube-api-access-fld2t\") pod \"barbican-db-create-gp4z9\" (UID: \"2c95dc54-6062-4f5f-8ffa-79cd1a23749b\") " pod="openstack/barbican-db-create-gp4z9" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.612656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7lzm\" (UniqueName: \"kubernetes.io/projected/6d7796c3-c17f-4795-b874-d5196d0deab2-kube-api-access-z7lzm\") pod \"keystone-db-sync-978m6\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.653792 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gzvqb" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.680712 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gp4z9" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.770178 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.911080 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerStarted","Data":"224dbcb223e227fbba9cd303b145f9619c8aa4605a720c0d2ebbb675ceb3f2c2"} Oct 08 15:33:07 crc kubenswrapper[4945]: I1008 15:33:07.970217 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.970201313 podStartE2EDuration="16.970201313s" podCreationTimestamp="2025-10-08 15:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:33:07.968406138 +0000 UTC m=+1197.322321039" watchObservedRunningTime="2025-10-08 15:33:07.970201313 +0000 UTC m=+1197.324116214" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.312534 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-gzvqb"] Oct 08 15:33:08 crc kubenswrapper[4945]: W1008 15:33:08.312909 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod549f6009_8c9d_4fbb_890d_236b5c93c3fa.slice/crio-2a51139c9e31ba3693685558dc853e1fef5fa15f6a280dcf59ceb83ba4c2ecc7 WatchSource:0}: Error finding container 2a51139c9e31ba3693685558dc853e1fef5fa15f6a280dcf59ceb83ba4c2ecc7: Status 404 returned error can't find the container with id 2a51139c9e31ba3693685558dc853e1fef5fa15f6a280dcf59ceb83ba4c2ecc7 Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.394583 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gp4z9"] Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.408073 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-978m6"] Oct 08 15:33:08 crc kubenswrapper[4945]: W1008 15:33:08.422453 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d7796c3_c17f_4795_b874_d5196d0deab2.slice/crio-bac1e53dfb52037770eee248a3f1b0b08d301de52e39d7ba30b47418a497d5af WatchSource:0}: Error finding container bac1e53dfb52037770eee248a3f1b0b08d301de52e39d7ba30b47418a497d5af: Status 404 returned error can't find the container with id bac1e53dfb52037770eee248a3f1b0b08d301de52e39d7ba30b47418a497d5af Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.833550 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-42dgd"] Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.835209 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-42dgd" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.846172 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-42dgd"] Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.867786 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-t8trl"] Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.869267 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.871502 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-gtdt6" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.871530 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.873857 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-t8trl"] Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.923293 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gp4z9" event={"ID":"2c95dc54-6062-4f5f-8ffa-79cd1a23749b","Type":"ContainerStarted","Data":"bddda164933162b2371f0930e6e673b2c885adf33d9cb1e1876d9e9843580c57"} Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.923330 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gp4z9" event={"ID":"2c95dc54-6062-4f5f-8ffa-79cd1a23749b","Type":"ContainerStarted","Data":"f3d61bc22add2a8ca40a9f2ea56c6b7220179c8f072ca7bc65ecc85ce4bf36e9"} Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.926436 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-978m6" event={"ID":"6d7796c3-c17f-4795-b874-d5196d0deab2","Type":"ContainerStarted","Data":"bac1e53dfb52037770eee248a3f1b0b08d301de52e39d7ba30b47418a497d5af"} Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.929810 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-gzvqb" event={"ID":"549f6009-8c9d-4fbb-890d-236b5c93c3fa","Type":"ContainerStarted","Data":"48df158afb6810f473c1638e7e7ae4e866fc706660c783a9be5d2ad845c7003f"} Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.929854 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-gzvqb" event={"ID":"549f6009-8c9d-4fbb-890d-236b5c93c3fa","Type":"ContainerStarted","Data":"2a51139c9e31ba3693685558dc853e1fef5fa15f6a280dcf59ceb83ba4c2ecc7"} Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.934898 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrq8z\" (UniqueName: \"kubernetes.io/projected/09e2edcd-fb7f-4c8b-aac6-7c6c2769e579-kube-api-access-nrq8z\") pod \"glance-db-create-42dgd\" (UID: \"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579\") " pod="openstack/glance-db-create-42dgd" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.934962 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-config-data\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.934994 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx6vj\" (UniqueName: \"kubernetes.io/projected/cf47e507-de36-49a3-9428-7104713b4d54-kube-api-access-bx6vj\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.935049 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-combined-ca-bundle\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.935169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-db-sync-config-data\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.942563 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-gp4z9" podStartSLOduration=1.942546409 podStartE2EDuration="1.942546409s" podCreationTimestamp="2025-10-08 15:33:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:33:08.936617949 +0000 UTC m=+1198.290532850" watchObservedRunningTime="2025-10-08 15:33:08.942546409 +0000 UTC m=+1198.296461310" Oct 08 15:33:08 crc kubenswrapper[4945]: I1008 15:33:08.964468 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-gzvqb" podStartSLOduration=1.964434078 podStartE2EDuration="1.964434078s" podCreationTimestamp="2025-10-08 15:33:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:33:08.959075103 +0000 UTC m=+1198.312989994" watchObservedRunningTime="2025-10-08 15:33:08.964434078 +0000 UTC m=+1198.318348979" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.028045 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-wlvmx"] Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.029439 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wlvmx" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.033936 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wlvmx"] Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.036658 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrq8z\" (UniqueName: \"kubernetes.io/projected/09e2edcd-fb7f-4c8b-aac6-7c6c2769e579-kube-api-access-nrq8z\") pod \"glance-db-create-42dgd\" (UID: \"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579\") " pod="openstack/glance-db-create-42dgd" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.036697 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-config-data\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.036719 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx6vj\" (UniqueName: \"kubernetes.io/projected/cf47e507-de36-49a3-9428-7104713b4d54-kube-api-access-bx6vj\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.036776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-combined-ca-bundle\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.036893 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-db-sync-config-data\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.044860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-combined-ca-bundle\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.044926 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-config-data\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.047514 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-db-sync-config-data\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.058847 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrq8z\" (UniqueName: \"kubernetes.io/projected/09e2edcd-fb7f-4c8b-aac6-7c6c2769e579-kube-api-access-nrq8z\") pod \"glance-db-create-42dgd\" (UID: \"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579\") " pod="openstack/glance-db-create-42dgd" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.059036 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx6vj\" (UniqueName: \"kubernetes.io/projected/cf47e507-de36-49a3-9428-7104713b4d54-kube-api-access-bx6vj\") pod \"watcher-db-sync-t8trl\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.139053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jxpw\" (UniqueName: \"kubernetes.io/projected/79250da1-c44a-4d62-a5e8-1ce671a336a2-kube-api-access-2jxpw\") pod \"neutron-db-create-wlvmx\" (UID: \"79250da1-c44a-4d62-a5e8-1ce671a336a2\") " pod="openstack/neutron-db-create-wlvmx" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.153902 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-42dgd" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.186892 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-t8trl" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.241245 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jxpw\" (UniqueName: \"kubernetes.io/projected/79250da1-c44a-4d62-a5e8-1ce671a336a2-kube-api-access-2jxpw\") pod \"neutron-db-create-wlvmx\" (UID: \"79250da1-c44a-4d62-a5e8-1ce671a336a2\") " pod="openstack/neutron-db-create-wlvmx" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.266806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jxpw\" (UniqueName: \"kubernetes.io/projected/79250da1-c44a-4d62-a5e8-1ce671a336a2-kube-api-access-2jxpw\") pod \"neutron-db-create-wlvmx\" (UID: \"79250da1-c44a-4d62-a5e8-1ce671a336a2\") " pod="openstack/neutron-db-create-wlvmx" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.417397 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wlvmx" Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.615677 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-42dgd"] Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.724900 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-t8trl"] Oct 08 15:33:09 crc kubenswrapper[4945]: W1008 15:33:09.738698 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf47e507_de36_49a3_9428_7104713b4d54.slice/crio-94b44fd5d6163d59aa5a81bbd4f9ce57ac70f714549c8003e98b01aab714588a WatchSource:0}: Error finding container 94b44fd5d6163d59aa5a81bbd4f9ce57ac70f714549c8003e98b01aab714588a: Status 404 returned error can't find the container with id 94b44fd5d6163d59aa5a81bbd4f9ce57ac70f714549c8003e98b01aab714588a Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.875090 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wlvmx"] Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.939836 4945 generic.go:334] "Generic (PLEG): container finished" podID="549f6009-8c9d-4fbb-890d-236b5c93c3fa" containerID="48df158afb6810f473c1638e7e7ae4e866fc706660c783a9be5d2ad845c7003f" exitCode=0 Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.939968 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-gzvqb" event={"ID":"549f6009-8c9d-4fbb-890d-236b5c93c3fa","Type":"ContainerDied","Data":"48df158afb6810f473c1638e7e7ae4e866fc706660c783a9be5d2ad845c7003f"} Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.941482 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wlvmx" event={"ID":"79250da1-c44a-4d62-a5e8-1ce671a336a2","Type":"ContainerStarted","Data":"d77744f14367a648fa32236fd8f9a06303f11806fd869d4bac8c3ef70d66a98d"} Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.943757 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-42dgd" event={"ID":"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579","Type":"ContainerStarted","Data":"aa1eb389650b5be13d90a8c5e975c9718a381bd539f4a8ed49626d1e404fef5e"} Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.944083 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-42dgd" event={"ID":"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579","Type":"ContainerStarted","Data":"ce92359997024bf4d3d57f15b28533004c70da34f2302e9547b414f9fe07f3f3"} Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.945581 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-t8trl" event={"ID":"cf47e507-de36-49a3-9428-7104713b4d54","Type":"ContainerStarted","Data":"94b44fd5d6163d59aa5a81bbd4f9ce57ac70f714549c8003e98b01aab714588a"} Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.947831 4945 generic.go:334] "Generic (PLEG): container finished" podID="2c95dc54-6062-4f5f-8ffa-79cd1a23749b" containerID="bddda164933162b2371f0930e6e673b2c885adf33d9cb1e1876d9e9843580c57" exitCode=0 Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.947859 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gp4z9" event={"ID":"2c95dc54-6062-4f5f-8ffa-79cd1a23749b","Type":"ContainerDied","Data":"bddda164933162b2371f0930e6e673b2c885adf33d9cb1e1876d9e9843580c57"} Oct 08 15:33:09 crc kubenswrapper[4945]: I1008 15:33:09.975456 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-42dgd" podStartSLOduration=1.975438054 podStartE2EDuration="1.975438054s" podCreationTimestamp="2025-10-08 15:33:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:33:09.965390162 +0000 UTC m=+1199.319305063" watchObservedRunningTime="2025-10-08 15:33:09.975438054 +0000 UTC m=+1199.329352955" Oct 08 15:33:10 crc kubenswrapper[4945]: I1008 15:33:10.966435 4945 generic.go:334] "Generic (PLEG): container finished" podID="09e2edcd-fb7f-4c8b-aac6-7c6c2769e579" containerID="aa1eb389650b5be13d90a8c5e975c9718a381bd539f4a8ed49626d1e404fef5e" exitCode=0 Oct 08 15:33:10 crc kubenswrapper[4945]: I1008 15:33:10.966497 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-42dgd" event={"ID":"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579","Type":"ContainerDied","Data":"aa1eb389650b5be13d90a8c5e975c9718a381bd539f4a8ed49626d1e404fef5e"} Oct 08 15:33:10 crc kubenswrapper[4945]: I1008 15:33:10.971071 4945 generic.go:334] "Generic (PLEG): container finished" podID="79250da1-c44a-4d62-a5e8-1ce671a336a2" containerID="4f845e6f03da20b18248928b9a4c30725d8c04becebef9b60f7c42e2475823ac" exitCode=0 Oct 08 15:33:10 crc kubenswrapper[4945]: I1008 15:33:10.971255 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wlvmx" event={"ID":"79250da1-c44a-4d62-a5e8-1ce671a336a2","Type":"ContainerDied","Data":"4f845e6f03da20b18248928b9a4c30725d8c04becebef9b60f7c42e2475823ac"} Oct 08 15:33:12 crc kubenswrapper[4945]: I1008 15:33:12.435937 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 15:33:14 crc kubenswrapper[4945]: I1008 15:33:14.916479 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wlvmx" Oct 08 15:33:14 crc kubenswrapper[4945]: I1008 15:33:14.923179 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gzvqb" Oct 08 15:33:14 crc kubenswrapper[4945]: I1008 15:33:14.931762 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gp4z9" Oct 08 15:33:14 crc kubenswrapper[4945]: I1008 15:33:14.940632 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-42dgd" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.011409 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-42dgd" event={"ID":"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579","Type":"ContainerDied","Data":"ce92359997024bf4d3d57f15b28533004c70da34f2302e9547b414f9fe07f3f3"} Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.011496 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce92359997024bf4d3d57f15b28533004c70da34f2302e9547b414f9fe07f3f3" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.011842 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-42dgd" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.013372 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gp4z9" event={"ID":"2c95dc54-6062-4f5f-8ffa-79cd1a23749b","Type":"ContainerDied","Data":"f3d61bc22add2a8ca40a9f2ea56c6b7220179c8f072ca7bc65ecc85ce4bf36e9"} Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.013403 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3d61bc22add2a8ca40a9f2ea56c6b7220179c8f072ca7bc65ecc85ce4bf36e9" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.013460 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gp4z9" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.015676 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gzvqb" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.015767 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-gzvqb" event={"ID":"549f6009-8c9d-4fbb-890d-236b5c93c3fa","Type":"ContainerDied","Data":"2a51139c9e31ba3693685558dc853e1fef5fa15f6a280dcf59ceb83ba4c2ecc7"} Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.015808 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a51139c9e31ba3693685558dc853e1fef5fa15f6a280dcf59ceb83ba4c2ecc7" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.017659 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wlvmx" event={"ID":"79250da1-c44a-4d62-a5e8-1ce671a336a2","Type":"ContainerDied","Data":"d77744f14367a648fa32236fd8f9a06303f11806fd869d4bac8c3ef70d66a98d"} Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.017694 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d77744f14367a648fa32236fd8f9a06303f11806fd869d4bac8c3ef70d66a98d" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.017702 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wlvmx" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.051348 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jxpw\" (UniqueName: \"kubernetes.io/projected/79250da1-c44a-4d62-a5e8-1ce671a336a2-kube-api-access-2jxpw\") pod \"79250da1-c44a-4d62-a5e8-1ce671a336a2\" (UID: \"79250da1-c44a-4d62-a5e8-1ce671a336a2\") " Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.051743 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fld2t\" (UniqueName: \"kubernetes.io/projected/2c95dc54-6062-4f5f-8ffa-79cd1a23749b-kube-api-access-fld2t\") pod \"2c95dc54-6062-4f5f-8ffa-79cd1a23749b\" (UID: \"2c95dc54-6062-4f5f-8ffa-79cd1a23749b\") " Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.051954 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrq8z\" (UniqueName: \"kubernetes.io/projected/09e2edcd-fb7f-4c8b-aac6-7c6c2769e579-kube-api-access-nrq8z\") pod \"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579\" (UID: \"09e2edcd-fb7f-4c8b-aac6-7c6c2769e579\") " Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.052061 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gmcx\" (UniqueName: \"kubernetes.io/projected/549f6009-8c9d-4fbb-890d-236b5c93c3fa-kube-api-access-9gmcx\") pod \"549f6009-8c9d-4fbb-890d-236b5c93c3fa\" (UID: \"549f6009-8c9d-4fbb-890d-236b5c93c3fa\") " Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.058611 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79250da1-c44a-4d62-a5e8-1ce671a336a2-kube-api-access-2jxpw" (OuterVolumeSpecName: "kube-api-access-2jxpw") pod "79250da1-c44a-4d62-a5e8-1ce671a336a2" (UID: "79250da1-c44a-4d62-a5e8-1ce671a336a2"). InnerVolumeSpecName "kube-api-access-2jxpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.064836 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549f6009-8c9d-4fbb-890d-236b5c93c3fa-kube-api-access-9gmcx" (OuterVolumeSpecName: "kube-api-access-9gmcx") pod "549f6009-8c9d-4fbb-890d-236b5c93c3fa" (UID: "549f6009-8c9d-4fbb-890d-236b5c93c3fa"). InnerVolumeSpecName "kube-api-access-9gmcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.064931 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e2edcd-fb7f-4c8b-aac6-7c6c2769e579-kube-api-access-nrq8z" (OuterVolumeSpecName: "kube-api-access-nrq8z") pod "09e2edcd-fb7f-4c8b-aac6-7c6c2769e579" (UID: "09e2edcd-fb7f-4c8b-aac6-7c6c2769e579"). InnerVolumeSpecName "kube-api-access-nrq8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.065174 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c95dc54-6062-4f5f-8ffa-79cd1a23749b-kube-api-access-fld2t" (OuterVolumeSpecName: "kube-api-access-fld2t") pod "2c95dc54-6062-4f5f-8ffa-79cd1a23749b" (UID: "2c95dc54-6062-4f5f-8ffa-79cd1a23749b"). InnerVolumeSpecName "kube-api-access-fld2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.154829 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gmcx\" (UniqueName: \"kubernetes.io/projected/549f6009-8c9d-4fbb-890d-236b5c93c3fa-kube-api-access-9gmcx\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.154893 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jxpw\" (UniqueName: \"kubernetes.io/projected/79250da1-c44a-4d62-a5e8-1ce671a336a2-kube-api-access-2jxpw\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.154921 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fld2t\" (UniqueName: \"kubernetes.io/projected/2c95dc54-6062-4f5f-8ffa-79cd1a23749b-kube-api-access-fld2t\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:15 crc kubenswrapper[4945]: I1008 15:33:15.154947 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrq8z\" (UniqueName: \"kubernetes.io/projected/09e2edcd-fb7f-4c8b-aac6-7c6c2769e579-kube-api-access-nrq8z\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:19 crc kubenswrapper[4945]: I1008 15:33:19.184374 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:33:19 crc kubenswrapper[4945]: I1008 15:33:19.185253 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:33:19 crc kubenswrapper[4945]: I1008 15:33:19.185335 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:33:19 crc kubenswrapper[4945]: I1008 15:33:19.186320 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d35c98890277031b010217d13608d3edf99fe1fc8eaff9d1b2452633ef6288de"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:33:19 crc kubenswrapper[4945]: I1008 15:33:19.186447 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://d35c98890277031b010217d13608d3edf99fe1fc8eaff9d1b2452633ef6288de" gracePeriod=600 Oct 08 15:33:20 crc kubenswrapper[4945]: I1008 15:33:20.063532 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="d35c98890277031b010217d13608d3edf99fe1fc8eaff9d1b2452633ef6288de" exitCode=0 Oct 08 15:33:20 crc kubenswrapper[4945]: I1008 15:33:20.063627 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"d35c98890277031b010217d13608d3edf99fe1fc8eaff9d1b2452633ef6288de"} Oct 08 15:33:20 crc kubenswrapper[4945]: I1008 15:33:20.063909 4945 scope.go:117] "RemoveContainer" containerID="06c6bd45fc6832af457ec9bd8f12ba1218de606e23dd6ce0bb0b843b24d0f361" Oct 08 15:33:22 crc kubenswrapper[4945]: I1008 15:33:22.435150 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 15:33:22 crc kubenswrapper[4945]: I1008 15:33:22.440719 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 15:33:23 crc kubenswrapper[4945]: I1008 15:33:23.106099 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.259595 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e695-account-create-tnzkz"] Oct 08 15:33:27 crc kubenswrapper[4945]: E1008 15:33:27.260652 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549f6009-8c9d-4fbb-890d-236b5c93c3fa" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.260669 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="549f6009-8c9d-4fbb-890d-236b5c93c3fa" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: E1008 15:33:27.260684 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e2edcd-fb7f-4c8b-aac6-7c6c2769e579" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.260692 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e2edcd-fb7f-4c8b-aac6-7c6c2769e579" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: E1008 15:33:27.260712 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79250da1-c44a-4d62-a5e8-1ce671a336a2" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.260721 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="79250da1-c44a-4d62-a5e8-1ce671a336a2" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: E1008 15:33:27.260735 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c95dc54-6062-4f5f-8ffa-79cd1a23749b" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.260745 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c95dc54-6062-4f5f-8ffa-79cd1a23749b" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.260997 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="79250da1-c44a-4d62-a5e8-1ce671a336a2" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.261016 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e2edcd-fb7f-4c8b-aac6-7c6c2769e579" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.261038 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c95dc54-6062-4f5f-8ffa-79cd1a23749b" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.261052 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="549f6009-8c9d-4fbb-890d-236b5c93c3fa" containerName="mariadb-database-create" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.261836 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e695-account-create-tnzkz" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.264444 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.268800 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e695-account-create-tnzkz"] Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.351384 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-80a1-account-create-gzqfc"] Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.352494 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-80a1-account-create-gzqfc" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.354899 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.362577 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4j4c\" (UniqueName: \"kubernetes.io/projected/e9316cb7-a351-4ca8-8ea9-e91fd507ea18-kube-api-access-m4j4c\") pod \"barbican-e695-account-create-tnzkz\" (UID: \"e9316cb7-a351-4ca8-8ea9-e91fd507ea18\") " pod="openstack/barbican-e695-account-create-tnzkz" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.363847 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-80a1-account-create-gzqfc"] Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.464978 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bhlh\" (UniqueName: \"kubernetes.io/projected/a113f9df-ecb0-48a1-9b10-05df2fea24c7-kube-api-access-9bhlh\") pod \"cinder-80a1-account-create-gzqfc\" (UID: \"a113f9df-ecb0-48a1-9b10-05df2fea24c7\") " pod="openstack/cinder-80a1-account-create-gzqfc" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.465047 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4j4c\" (UniqueName: \"kubernetes.io/projected/e9316cb7-a351-4ca8-8ea9-e91fd507ea18-kube-api-access-m4j4c\") pod \"barbican-e695-account-create-tnzkz\" (UID: \"e9316cb7-a351-4ca8-8ea9-e91fd507ea18\") " pod="openstack/barbican-e695-account-create-tnzkz" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.484030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4j4c\" (UniqueName: \"kubernetes.io/projected/e9316cb7-a351-4ca8-8ea9-e91fd507ea18-kube-api-access-m4j4c\") pod \"barbican-e695-account-create-tnzkz\" (UID: \"e9316cb7-a351-4ca8-8ea9-e91fd507ea18\") " pod="openstack/barbican-e695-account-create-tnzkz" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.566300 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bhlh\" (UniqueName: \"kubernetes.io/projected/a113f9df-ecb0-48a1-9b10-05df2fea24c7-kube-api-access-9bhlh\") pod \"cinder-80a1-account-create-gzqfc\" (UID: \"a113f9df-ecb0-48a1-9b10-05df2fea24c7\") " pod="openstack/cinder-80a1-account-create-gzqfc" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.582208 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bhlh\" (UniqueName: \"kubernetes.io/projected/a113f9df-ecb0-48a1-9b10-05df2fea24c7-kube-api-access-9bhlh\") pod \"cinder-80a1-account-create-gzqfc\" (UID: \"a113f9df-ecb0-48a1-9b10-05df2fea24c7\") " pod="openstack/cinder-80a1-account-create-gzqfc" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.583285 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e695-account-create-tnzkz" Oct 08 15:33:27 crc kubenswrapper[4945]: I1008 15:33:27.675918 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-80a1-account-create-gzqfc" Oct 08 15:33:28 crc kubenswrapper[4945]: I1008 15:33:28.964482 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-1ef5-account-create-zz9f8"] Oct 08 15:33:28 crc kubenswrapper[4945]: I1008 15:33:28.966221 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1ef5-account-create-zz9f8" Oct 08 15:33:28 crc kubenswrapper[4945]: I1008 15:33:28.970845 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 08 15:33:28 crc kubenswrapper[4945]: I1008 15:33:28.978436 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1ef5-account-create-zz9f8"] Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.091253 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdmjl\" (UniqueName: \"kubernetes.io/projected/0462b93b-24b8-42a6-81cf-30e0b0b09eb7-kube-api-access-rdmjl\") pod \"glance-1ef5-account-create-zz9f8\" (UID: \"0462b93b-24b8-42a6-81cf-30e0b0b09eb7\") " pod="openstack/glance-1ef5-account-create-zz9f8" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.168424 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-37c5-account-create-dwc7q"] Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.170020 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-37c5-account-create-dwc7q" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.173263 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.175771 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-37c5-account-create-dwc7q"] Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.192636 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdmjl\" (UniqueName: \"kubernetes.io/projected/0462b93b-24b8-42a6-81cf-30e0b0b09eb7-kube-api-access-rdmjl\") pod \"glance-1ef5-account-create-zz9f8\" (UID: \"0462b93b-24b8-42a6-81cf-30e0b0b09eb7\") " pod="openstack/glance-1ef5-account-create-zz9f8" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.236901 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdmjl\" (UniqueName: \"kubernetes.io/projected/0462b93b-24b8-42a6-81cf-30e0b0b09eb7-kube-api-access-rdmjl\") pod \"glance-1ef5-account-create-zz9f8\" (UID: \"0462b93b-24b8-42a6-81cf-30e0b0b09eb7\") " pod="openstack/glance-1ef5-account-create-zz9f8" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.294019 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86kqn\" (UniqueName: \"kubernetes.io/projected/269baca3-654a-415a-a631-258a2bc03e03-kube-api-access-86kqn\") pod \"neutron-37c5-account-create-dwc7q\" (UID: \"269baca3-654a-415a-a631-258a2bc03e03\") " pod="openstack/neutron-37c5-account-create-dwc7q" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.302229 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1ef5-account-create-zz9f8" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.395389 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86kqn\" (UniqueName: \"kubernetes.io/projected/269baca3-654a-415a-a631-258a2bc03e03-kube-api-access-86kqn\") pod \"neutron-37c5-account-create-dwc7q\" (UID: \"269baca3-654a-415a-a631-258a2bc03e03\") " pod="openstack/neutron-37c5-account-create-dwc7q" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.415784 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86kqn\" (UniqueName: \"kubernetes.io/projected/269baca3-654a-415a-a631-258a2bc03e03-kube-api-access-86kqn\") pod \"neutron-37c5-account-create-dwc7q\" (UID: \"269baca3-654a-415a-a631-258a2bc03e03\") " pod="openstack/neutron-37c5-account-create-dwc7q" Oct 08 15:33:29 crc kubenswrapper[4945]: I1008 15:33:29.492821 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-37c5-account-create-dwc7q" Oct 08 15:33:31 crc kubenswrapper[4945]: E1008 15:33:31.738187 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Oct 08 15:33:31 crc kubenswrapper[4945]: E1008 15:33:31.738474 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Oct 08 15:33:31 crc kubenswrapper[4945]: E1008 15:33:31.738632 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:watcher-db-sync,Image:38.102.83.163:5001/podified-master-centos10/openstack-watcher-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/watcher/watcher.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:watcher-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bx6vj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-db-sync-t8trl_openstack(cf47e507-de36-49a3-9428-7104713b4d54): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:33:31 crc kubenswrapper[4945]: E1008 15:33:31.741031 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/watcher-db-sync-t8trl" podUID="cf47e507-de36-49a3-9428-7104713b4d54" Oct 08 15:33:32 crc kubenswrapper[4945]: I1008 15:33:32.187751 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"d5311a5a778713e57615c5afa0552244f2b5dba47b435a164e79654109f1ba68"} Oct 08 15:33:32 crc kubenswrapper[4945]: I1008 15:33:32.207573 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-978m6" event={"ID":"6d7796c3-c17f-4795-b874-d5196d0deab2","Type":"ContainerStarted","Data":"8b220812c914192ec6cd78a4b415bb6f300201396cff210462ed3c4336553bed"} Oct 08 15:33:32 crc kubenswrapper[4945]: E1008 15:33:32.214860 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-watcher-api:watcher_latest\\\"\"" pod="openstack/watcher-db-sync-t8trl" podUID="cf47e507-de36-49a3-9428-7104713b4d54" Oct 08 15:33:32 crc kubenswrapper[4945]: I1008 15:33:32.227179 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-80a1-account-create-gzqfc"] Oct 08 15:33:32 crc kubenswrapper[4945]: I1008 15:33:32.242354 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-978m6" podStartSLOduration=1.936982108 podStartE2EDuration="25.24233648s" podCreationTimestamp="2025-10-08 15:33:07 +0000 UTC" firstStartedPulling="2025-10-08 15:33:08.431470025 +0000 UTC m=+1197.785384916" lastFinishedPulling="2025-10-08 15:33:31.736824387 +0000 UTC m=+1221.090739288" observedRunningTime="2025-10-08 15:33:32.230091042 +0000 UTC m=+1221.584005933" watchObservedRunningTime="2025-10-08 15:33:32.24233648 +0000 UTC m=+1221.596251381" Oct 08 15:33:33 crc kubenswrapper[4945]: W1008 15:33:33.173684 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod269baca3_654a_415a_a631_258a2bc03e03.slice/crio-412eb35e9ab3a26a0ee69b98d39577142d2df960aaa8bf5b29434bf2294ddc14 WatchSource:0}: Error finding container 412eb35e9ab3a26a0ee69b98d39577142d2df960aaa8bf5b29434bf2294ddc14: Status 404 returned error can't find the container with id 412eb35e9ab3a26a0ee69b98d39577142d2df960aaa8bf5b29434bf2294ddc14 Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.181176 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-37c5-account-create-dwc7q"] Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.189135 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e695-account-create-tnzkz"] Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.215747 4945 generic.go:334] "Generic (PLEG): container finished" podID="a113f9df-ecb0-48a1-9b10-05df2fea24c7" containerID="6ec53bed8a5ddd4fbd36c6e1b8be7a4f7e4b9a6612f1176c25c5f524b6f1110a" exitCode=0 Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.215824 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-80a1-account-create-gzqfc" event={"ID":"a113f9df-ecb0-48a1-9b10-05df2fea24c7","Type":"ContainerDied","Data":"6ec53bed8a5ddd4fbd36c6e1b8be7a4f7e4b9a6612f1176c25c5f524b6f1110a"} Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.215852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-80a1-account-create-gzqfc" event={"ID":"a113f9df-ecb0-48a1-9b10-05df2fea24c7","Type":"ContainerStarted","Data":"6388d34112dc4a89110db9181bb125cb7a04df7bb82652f3bcbee0718162c15a"} Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.216981 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e695-account-create-tnzkz" event={"ID":"e9316cb7-a351-4ca8-8ea9-e91fd507ea18","Type":"ContainerStarted","Data":"bee130aeffd3547c62ecdcddc3eec8214a04c4b5a1195afec46d49287a3a5494"} Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.219241 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-37c5-account-create-dwc7q" event={"ID":"269baca3-654a-415a-a631-258a2bc03e03","Type":"ContainerStarted","Data":"412eb35e9ab3a26a0ee69b98d39577142d2df960aaa8bf5b29434bf2294ddc14"} Oct 08 15:33:33 crc kubenswrapper[4945]: I1008 15:33:33.283447 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1ef5-account-create-zz9f8"] Oct 08 15:33:33 crc kubenswrapper[4945]: W1008 15:33:33.285789 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0462b93b_24b8_42a6_81cf_30e0b0b09eb7.slice/crio-3a9a93b3815b149fff24383482356989e46dd445abeb4b3c1b3b640567e349bf WatchSource:0}: Error finding container 3a9a93b3815b149fff24383482356989e46dd445abeb4b3c1b3b640567e349bf: Status 404 returned error can't find the container with id 3a9a93b3815b149fff24383482356989e46dd445abeb4b3c1b3b640567e349bf Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.229424 4945 generic.go:334] "Generic (PLEG): container finished" podID="0462b93b-24b8-42a6-81cf-30e0b0b09eb7" containerID="0ccf493da6ed3153d0181e26501441b476f23b3c48a6eb8d520752e3766b822f" exitCode=0 Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.229541 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1ef5-account-create-zz9f8" event={"ID":"0462b93b-24b8-42a6-81cf-30e0b0b09eb7","Type":"ContainerDied","Data":"0ccf493da6ed3153d0181e26501441b476f23b3c48a6eb8d520752e3766b822f"} Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.229787 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1ef5-account-create-zz9f8" event={"ID":"0462b93b-24b8-42a6-81cf-30e0b0b09eb7","Type":"ContainerStarted","Data":"3a9a93b3815b149fff24383482356989e46dd445abeb4b3c1b3b640567e349bf"} Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.232231 4945 generic.go:334] "Generic (PLEG): container finished" podID="e9316cb7-a351-4ca8-8ea9-e91fd507ea18" containerID="d05d6192f8559971e4cf7034bd5fd3505c6dc3343a8d581776f7a9847d8aa77a" exitCode=0 Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.232321 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e695-account-create-tnzkz" event={"ID":"e9316cb7-a351-4ca8-8ea9-e91fd507ea18","Type":"ContainerDied","Data":"d05d6192f8559971e4cf7034bd5fd3505c6dc3343a8d581776f7a9847d8aa77a"} Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.234262 4945 generic.go:334] "Generic (PLEG): container finished" podID="269baca3-654a-415a-a631-258a2bc03e03" containerID="698be937fa5f0c9c26664708304be7fc79e96fc724ce9fee1cdbb51c8ab42c80" exitCode=0 Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.234295 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-37c5-account-create-dwc7q" event={"ID":"269baca3-654a-415a-a631-258a2bc03e03","Type":"ContainerDied","Data":"698be937fa5f0c9c26664708304be7fc79e96fc724ce9fee1cdbb51c8ab42c80"} Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.631986 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-80a1-account-create-gzqfc" Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.788763 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bhlh\" (UniqueName: \"kubernetes.io/projected/a113f9df-ecb0-48a1-9b10-05df2fea24c7-kube-api-access-9bhlh\") pod \"a113f9df-ecb0-48a1-9b10-05df2fea24c7\" (UID: \"a113f9df-ecb0-48a1-9b10-05df2fea24c7\") " Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.794783 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a113f9df-ecb0-48a1-9b10-05df2fea24c7-kube-api-access-9bhlh" (OuterVolumeSpecName: "kube-api-access-9bhlh") pod "a113f9df-ecb0-48a1-9b10-05df2fea24c7" (UID: "a113f9df-ecb0-48a1-9b10-05df2fea24c7"). InnerVolumeSpecName "kube-api-access-9bhlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:34 crc kubenswrapper[4945]: I1008 15:33:34.890835 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bhlh\" (UniqueName: \"kubernetes.io/projected/a113f9df-ecb0-48a1-9b10-05df2fea24c7-kube-api-access-9bhlh\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.249348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-80a1-account-create-gzqfc" event={"ID":"a113f9df-ecb0-48a1-9b10-05df2fea24c7","Type":"ContainerDied","Data":"6388d34112dc4a89110db9181bb125cb7a04df7bb82652f3bcbee0718162c15a"} Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.249850 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6388d34112dc4a89110db9181bb125cb7a04df7bb82652f3bcbee0718162c15a" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.249522 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-80a1-account-create-gzqfc" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.733769 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-37c5-account-create-dwc7q" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.742077 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1ef5-account-create-zz9f8" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.749054 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e695-account-create-tnzkz" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.812504 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdmjl\" (UniqueName: \"kubernetes.io/projected/0462b93b-24b8-42a6-81cf-30e0b0b09eb7-kube-api-access-rdmjl\") pod \"0462b93b-24b8-42a6-81cf-30e0b0b09eb7\" (UID: \"0462b93b-24b8-42a6-81cf-30e0b0b09eb7\") " Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.812591 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86kqn\" (UniqueName: \"kubernetes.io/projected/269baca3-654a-415a-a631-258a2bc03e03-kube-api-access-86kqn\") pod \"269baca3-654a-415a-a631-258a2bc03e03\" (UID: \"269baca3-654a-415a-a631-258a2bc03e03\") " Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.812628 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4j4c\" (UniqueName: \"kubernetes.io/projected/e9316cb7-a351-4ca8-8ea9-e91fd507ea18-kube-api-access-m4j4c\") pod \"e9316cb7-a351-4ca8-8ea9-e91fd507ea18\" (UID: \"e9316cb7-a351-4ca8-8ea9-e91fd507ea18\") " Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.819348 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9316cb7-a351-4ca8-8ea9-e91fd507ea18-kube-api-access-m4j4c" (OuterVolumeSpecName: "kube-api-access-m4j4c") pod "e9316cb7-a351-4ca8-8ea9-e91fd507ea18" (UID: "e9316cb7-a351-4ca8-8ea9-e91fd507ea18"). InnerVolumeSpecName "kube-api-access-m4j4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.820044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0462b93b-24b8-42a6-81cf-30e0b0b09eb7-kube-api-access-rdmjl" (OuterVolumeSpecName: "kube-api-access-rdmjl") pod "0462b93b-24b8-42a6-81cf-30e0b0b09eb7" (UID: "0462b93b-24b8-42a6-81cf-30e0b0b09eb7"). InnerVolumeSpecName "kube-api-access-rdmjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.821713 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/269baca3-654a-415a-a631-258a2bc03e03-kube-api-access-86kqn" (OuterVolumeSpecName: "kube-api-access-86kqn") pod "269baca3-654a-415a-a631-258a2bc03e03" (UID: "269baca3-654a-415a-a631-258a2bc03e03"). InnerVolumeSpecName "kube-api-access-86kqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.914730 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdmjl\" (UniqueName: \"kubernetes.io/projected/0462b93b-24b8-42a6-81cf-30e0b0b09eb7-kube-api-access-rdmjl\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.914989 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86kqn\" (UniqueName: \"kubernetes.io/projected/269baca3-654a-415a-a631-258a2bc03e03-kube-api-access-86kqn\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:35 crc kubenswrapper[4945]: I1008 15:33:35.914998 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4j4c\" (UniqueName: \"kubernetes.io/projected/e9316cb7-a351-4ca8-8ea9-e91fd507ea18-kube-api-access-m4j4c\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.261912 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1ef5-account-create-zz9f8" event={"ID":"0462b93b-24b8-42a6-81cf-30e0b0b09eb7","Type":"ContainerDied","Data":"3a9a93b3815b149fff24383482356989e46dd445abeb4b3c1b3b640567e349bf"} Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.261957 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a9a93b3815b149fff24383482356989e46dd445abeb4b3c1b3b640567e349bf" Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.261980 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1ef5-account-create-zz9f8" Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.266705 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e695-account-create-tnzkz" event={"ID":"e9316cb7-a351-4ca8-8ea9-e91fd507ea18","Type":"ContainerDied","Data":"bee130aeffd3547c62ecdcddc3eec8214a04c4b5a1195afec46d49287a3a5494"} Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.266759 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bee130aeffd3547c62ecdcddc3eec8214a04c4b5a1195afec46d49287a3a5494" Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.266822 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e695-account-create-tnzkz" Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.269795 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-37c5-account-create-dwc7q" event={"ID":"269baca3-654a-415a-a631-258a2bc03e03","Type":"ContainerDied","Data":"412eb35e9ab3a26a0ee69b98d39577142d2df960aaa8bf5b29434bf2294ddc14"} Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.269826 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="412eb35e9ab3a26a0ee69b98d39577142d2df960aaa8bf5b29434bf2294ddc14" Oct 08 15:33:36 crc kubenswrapper[4945]: I1008 15:33:36.269864 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-37c5-account-create-dwc7q" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.039866 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4bnbb"] Oct 08 15:33:39 crc kubenswrapper[4945]: E1008 15:33:39.040796 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269baca3-654a-415a-a631-258a2bc03e03" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.040811 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="269baca3-654a-415a-a631-258a2bc03e03" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: E1008 15:33:39.040844 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9316cb7-a351-4ca8-8ea9-e91fd507ea18" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.040853 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9316cb7-a351-4ca8-8ea9-e91fd507ea18" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: E1008 15:33:39.040873 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0462b93b-24b8-42a6-81cf-30e0b0b09eb7" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.040882 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0462b93b-24b8-42a6-81cf-30e0b0b09eb7" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: E1008 15:33:39.040893 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113f9df-ecb0-48a1-9b10-05df2fea24c7" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.040900 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113f9df-ecb0-48a1-9b10-05df2fea24c7" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.041131 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="269baca3-654a-415a-a631-258a2bc03e03" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.041150 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9316cb7-a351-4ca8-8ea9-e91fd507ea18" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.041162 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113f9df-ecb0-48a1-9b10-05df2fea24c7" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.041180 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0462b93b-24b8-42a6-81cf-30e0b0b09eb7" containerName="mariadb-account-create" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.041861 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.044052 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g97m6" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.044440 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.053738 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4bnbb"] Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.181962 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-db-sync-config-data\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.182020 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-config-data\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.182522 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr722\" (UniqueName: \"kubernetes.io/projected/0646f7b0-6abf-436c-976f-59ab119525ad-kube-api-access-qr722\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.182686 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-combined-ca-bundle\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.284063 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr722\" (UniqueName: \"kubernetes.io/projected/0646f7b0-6abf-436c-976f-59ab119525ad-kube-api-access-qr722\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.284194 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-combined-ca-bundle\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.284253 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-db-sync-config-data\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.284298 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-config-data\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.293572 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-config-data\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.294009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-db-sync-config-data\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.299638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-combined-ca-bundle\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.324765 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr722\" (UniqueName: \"kubernetes.io/projected/0646f7b0-6abf-436c-976f-59ab119525ad-kube-api-access-qr722\") pod \"glance-db-sync-4bnbb\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.371664 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4bnbb" Oct 08 15:33:39 crc kubenswrapper[4945]: I1008 15:33:39.909414 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4bnbb"] Oct 08 15:33:40 crc kubenswrapper[4945]: I1008 15:33:40.313479 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4bnbb" event={"ID":"0646f7b0-6abf-436c-976f-59ab119525ad","Type":"ContainerStarted","Data":"e70453d1d42cfd4f9c54d651399dd52f250f583dd4fcc0b0420deea266708cc6"} Oct 08 15:33:53 crc kubenswrapper[4945]: E1008 15:33:53.149051 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 15:33:53 crc kubenswrapper[4945]: E1008 15:33:53.149523 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 15:33:53 crc kubenswrapper[4945]: E1008 15:33:53.149643 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qr722,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-4bnbb_openstack(0646f7b0-6abf-436c-976f-59ab119525ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:33:53 crc kubenswrapper[4945]: E1008 15:33:53.150893 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-4bnbb" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" Oct 08 15:33:53 crc kubenswrapper[4945]: I1008 15:33:53.432304 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-t8trl" event={"ID":"cf47e507-de36-49a3-9428-7104713b4d54","Type":"ContainerStarted","Data":"b027c7df84247536b2b76ba5946c3242f6ae1576ffde04680aeaeff006508020"} Oct 08 15:33:53 crc kubenswrapper[4945]: E1008 15:33:53.433800 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest\\\"\"" pod="openstack/glance-db-sync-4bnbb" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" Oct 08 15:33:54 crc kubenswrapper[4945]: I1008 15:33:54.444547 4945 generic.go:334] "Generic (PLEG): container finished" podID="6d7796c3-c17f-4795-b874-d5196d0deab2" containerID="8b220812c914192ec6cd78a4b415bb6f300201396cff210462ed3c4336553bed" exitCode=0 Oct 08 15:33:54 crc kubenswrapper[4945]: I1008 15:33:54.444612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-978m6" event={"ID":"6d7796c3-c17f-4795-b874-d5196d0deab2","Type":"ContainerDied","Data":"8b220812c914192ec6cd78a4b415bb6f300201396cff210462ed3c4336553bed"} Oct 08 15:33:54 crc kubenswrapper[4945]: I1008 15:33:54.491008 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-t8trl" podStartSLOduration=3.075206471 podStartE2EDuration="46.490976355s" podCreationTimestamp="2025-10-08 15:33:08 +0000 UTC" firstStartedPulling="2025-10-08 15:33:09.755832816 +0000 UTC m=+1199.109747717" lastFinishedPulling="2025-10-08 15:33:53.1716027 +0000 UTC m=+1242.525517601" observedRunningTime="2025-10-08 15:33:54.482356629 +0000 UTC m=+1243.836271560" watchObservedRunningTime="2025-10-08 15:33:54.490976355 +0000 UTC m=+1243.844891286" Oct 08 15:33:55 crc kubenswrapper[4945]: I1008 15:33:55.824174 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.000794 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-config-data\") pod \"6d7796c3-c17f-4795-b874-d5196d0deab2\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.000903 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7lzm\" (UniqueName: \"kubernetes.io/projected/6d7796c3-c17f-4795-b874-d5196d0deab2-kube-api-access-z7lzm\") pod \"6d7796c3-c17f-4795-b874-d5196d0deab2\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.001037 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-combined-ca-bundle\") pod \"6d7796c3-c17f-4795-b874-d5196d0deab2\" (UID: \"6d7796c3-c17f-4795-b874-d5196d0deab2\") " Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.007956 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d7796c3-c17f-4795-b874-d5196d0deab2-kube-api-access-z7lzm" (OuterVolumeSpecName: "kube-api-access-z7lzm") pod "6d7796c3-c17f-4795-b874-d5196d0deab2" (UID: "6d7796c3-c17f-4795-b874-d5196d0deab2"). InnerVolumeSpecName "kube-api-access-z7lzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.029198 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d7796c3-c17f-4795-b874-d5196d0deab2" (UID: "6d7796c3-c17f-4795-b874-d5196d0deab2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.062421 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-config-data" (OuterVolumeSpecName: "config-data") pod "6d7796c3-c17f-4795-b874-d5196d0deab2" (UID: "6d7796c3-c17f-4795-b874-d5196d0deab2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.103273 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.103309 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d7796c3-c17f-4795-b874-d5196d0deab2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.103322 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7lzm\" (UniqueName: \"kubernetes.io/projected/6d7796c3-c17f-4795-b874-d5196d0deab2-kube-api-access-z7lzm\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.473505 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-978m6" event={"ID":"6d7796c3-c17f-4795-b874-d5196d0deab2","Type":"ContainerDied","Data":"bac1e53dfb52037770eee248a3f1b0b08d301de52e39d7ba30b47418a497d5af"} Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.473564 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bac1e53dfb52037770eee248a3f1b0b08d301de52e39d7ba30b47418a497d5af" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.473659 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-978m6" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.766563 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56bd54c5ff-jwcfl"] Oct 08 15:33:56 crc kubenswrapper[4945]: E1008 15:33:56.766931 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d7796c3-c17f-4795-b874-d5196d0deab2" containerName="keystone-db-sync" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.766948 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d7796c3-c17f-4795-b874-d5196d0deab2" containerName="keystone-db-sync" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.767131 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d7796c3-c17f-4795-b874-d5196d0deab2" containerName="keystone-db-sync" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.767964 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.784571 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rsfkt"] Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.787976 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.789813 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7rf" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.792221 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.792308 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.792316 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.797185 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rsfkt"] Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.810405 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56bd54c5ff-jwcfl"] Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.924870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-config-data\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-credential-keys\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-swift-storage-0\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925284 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h472z\" (UniqueName: \"kubernetes.io/projected/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-kube-api-access-h472z\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz5kc\" (UniqueName: \"kubernetes.io/projected/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-kube-api-access-tz5kc\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925346 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-fernet-keys\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-config\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925399 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-scripts\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925432 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-nb\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925456 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-svc\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925499 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-combined-ca-bundle\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.925542 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-sb\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.941459 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68559d9d9f-k46pc"] Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.942914 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.947548 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.947704 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-rbpsw" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.947851 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.947911 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 08 15:33:56 crc kubenswrapper[4945]: I1008 15:33:56.967643 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68559d9d9f-k46pc"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.003248 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-ztjf5"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.004669 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.011179 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.011298 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-45f87" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.011469 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.016230 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ztjf5"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.026952 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h472z\" (UniqueName: \"kubernetes.io/projected/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-kube-api-access-h472z\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027043 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz5kc\" (UniqueName: \"kubernetes.io/projected/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-kube-api-access-tz5kc\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027069 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-fernet-keys\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027092 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-config\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027147 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-scripts\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027182 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-nb\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027207 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-svc\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-combined-ca-bundle\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027283 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-sb\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027354 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-config-data\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027376 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-credential-keys\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.027402 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-swift-storage-0\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.029045 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-swift-storage-0\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.029584 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-sb\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.033944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-nb\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.034512 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-config\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.035714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-svc\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.047836 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-credential-keys\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.048610 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-scripts\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.049191 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-combined-ca-bundle\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.049777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-fernet-keys\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.082173 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h472z\" (UniqueName: \"kubernetes.io/projected/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-kube-api-access-h472z\") pod \"dnsmasq-dns-56bd54c5ff-jwcfl\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.082828 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-config-data\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.094814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz5kc\" (UniqueName: \"kubernetes.io/projected/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-kube-api-access-tz5kc\") pod \"keystone-bootstrap-rsfkt\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.103414 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.126156 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.128648 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.129174 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.134265 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsg9v\" (UniqueName: \"kubernetes.io/projected/3dec4ad0-549d-49a0-9749-86b21bbaf987-kube-api-access-tsg9v\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.153997 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfz87\" (UniqueName: \"kubernetes.io/projected/6a3b7ec6-9364-4d86-981e-e882eb997e41-kube-api-access-qfz87\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154296 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-config-data\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-config\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154664 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-scripts\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154803 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-combined-ca-bundle\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154891 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3b7ec6-9364-4d86-981e-e882eb997e41-logs\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154960 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a3b7ec6-9364-4d86-981e-e882eb997e41-horizon-secret-key\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154484 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.154570 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.177897 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.215173 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-pfbzj"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.216354 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.239198 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pfbzj"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.245040 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fkwqt" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.245206 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.245302 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.251100 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56bd54c5ff-jwcfl"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.255990 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-scripts\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256041 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-log-httpd\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256073 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-scripts\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsg9v\" (UniqueName: \"kubernetes.io/projected/3dec4ad0-549d-49a0-9749-86b21bbaf987-kube-api-access-tsg9v\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256179 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-config-data\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256266 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfz87\" (UniqueName: \"kubernetes.io/projected/6a3b7ec6-9364-4d86-981e-e882eb997e41-kube-api-access-qfz87\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256292 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-config-data\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256343 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256368 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-config\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256390 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1be99142-5236-4365-8a45-70847b54b883-etc-machine-id\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256417 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-combined-ca-bundle\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256442 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-scripts\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256473 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-config-data\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256491 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-combined-ca-bundle\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256506 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-run-httpd\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256527 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3b7ec6-9364-4d86-981e-e882eb997e41-logs\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256546 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a3b7ec6-9364-4d86-981e-e882eb997e41-horizon-secret-key\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmnck\" (UniqueName: \"kubernetes.io/projected/86db9194-a67b-4491-b0f1-d03bafb3e353-kube-api-access-qmnck\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256597 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256616 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pclt8\" (UniqueName: \"kubernetes.io/projected/1be99142-5236-4365-8a45-70847b54b883-kube-api-access-pclt8\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.256637 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-db-sync-config-data\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.258590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3b7ec6-9364-4d86-981e-e882eb997e41-logs\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.258765 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-config-data\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.259227 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-scripts\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.267806 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-s7fh4"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.268876 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.272896 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a3b7ec6-9364-4d86-981e-e882eb997e41-horizon-secret-key\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.276894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-config\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.279870 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-combined-ca-bundle\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.284972 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.285435 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.285593 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f57676dc-brk7t"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.285849 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tfz6r" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.288141 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.294356 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfz87\" (UniqueName: \"kubernetes.io/projected/6a3b7ec6-9364-4d86-981e-e882eb997e41-kube-api-access-qfz87\") pod \"horizon-68559d9d9f-k46pc\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.294522 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s7fh4"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.308986 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79c9bb54c7-hz8j2"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.311009 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.337320 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f57676dc-brk7t"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.343789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsg9v\" (UniqueName: \"kubernetes.io/projected/3dec4ad0-549d-49a0-9749-86b21bbaf987-kube-api-access-tsg9v\") pod \"neutron-db-sync-ztjf5\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358144 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42416335-c299-401c-b732-ae75dfa6a1d2-logs\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-config-data\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358207 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/42416335-c299-401c-b732-ae75dfa6a1d2-horizon-secret-key\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358225 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-run-httpd\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358242 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt4mk\" (UniqueName: \"kubernetes.io/projected/42416335-c299-401c-b732-ae75dfa6a1d2-kube-api-access-wt4mk\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358265 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-config-data\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358293 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmnck\" (UniqueName: \"kubernetes.io/projected/86db9194-a67b-4491-b0f1-d03bafb3e353-kube-api-access-qmnck\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358316 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-scripts\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358334 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358352 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pclt8\" (UniqueName: \"kubernetes.io/projected/1be99142-5236-4365-8a45-70847b54b883-kube-api-access-pclt8\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-config\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358384 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-scripts\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358405 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-db-sync-config-data\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-nb\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358447 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-scripts\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358462 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-svc\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358482 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-combined-ca-bundle\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358504 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-log-httpd\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358520 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-scripts\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhj94\" (UniqueName: \"kubernetes.io/projected/4cdde67a-2aad-4ef4-ae1b-db394a522f32-kube-api-access-hhj94\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358562 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-config-data\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358593 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47794116-db67-4451-858c-bb6fea64a21c-logs\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358626 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-config-data\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358650 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-swift-storage-0\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358676 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358710 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbnnp\" (UniqueName: \"kubernetes.io/projected/47794116-db67-4451-858c-bb6fea64a21c-kube-api-access-lbnnp\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1be99142-5236-4365-8a45-70847b54b883-etc-machine-id\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358755 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-sb\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-combined-ca-bundle\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.358961 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1be99142-5236-4365-8a45-70847b54b883-etc-machine-id\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.365209 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-run-httpd\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.365920 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-log-httpd\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.370348 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79c9bb54c7-hz8j2"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.372899 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-config-data\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.376522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-scripts\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.377143 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-config-data\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.378676 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.380729 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-combined-ca-bundle\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.388252 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-db-sync-config-data\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.390530 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.390999 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-scripts\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.406638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pclt8\" (UniqueName: \"kubernetes.io/projected/1be99142-5236-4365-8a45-70847b54b883-kube-api-access-pclt8\") pod \"cinder-db-sync-pfbzj\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.409086 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ttghn"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.410236 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.413395 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kqbsg" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.413534 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.413986 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmnck\" (UniqueName: \"kubernetes.io/projected/86db9194-a67b-4491-b0f1-d03bafb3e353-kube-api-access-qmnck\") pod \"ceilometer-0\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.475129 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-nb\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.475361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-svc\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.485728 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-nb\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.487915 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-combined-ca-bundle\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.487990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhj94\" (UniqueName: \"kubernetes.io/projected/4cdde67a-2aad-4ef4-ae1b-db394a522f32-kube-api-access-hhj94\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488038 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47794116-db67-4451-858c-bb6fea64a21c-logs\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488097 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-config-data\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-swift-storage-0\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488243 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbnnp\" (UniqueName: \"kubernetes.io/projected/47794116-db67-4451-858c-bb6fea64a21c-kube-api-access-lbnnp\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488286 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-sb\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488348 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42416335-c299-401c-b732-ae75dfa6a1d2-logs\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488382 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/42416335-c299-401c-b732-ae75dfa6a1d2-horizon-secret-key\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488404 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt4mk\" (UniqueName: \"kubernetes.io/projected/42416335-c299-401c-b732-ae75dfa6a1d2-kube-api-access-wt4mk\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488454 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-config-data\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488514 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-scripts\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488547 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-config\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.488568 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-scripts\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.489873 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-svc\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.490350 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47794116-db67-4451-858c-bb6fea64a21c-logs\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.490631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42416335-c299-401c-b732-ae75dfa6a1d2-logs\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.491144 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-config-data\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.492757 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-sb\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.495290 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-scripts\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.495621 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-config\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.497701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-combined-ca-bundle\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.497554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-swift-storage-0\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.515300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-scripts\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.543380 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-config-data\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.543942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/42416335-c299-401c-b732-ae75dfa6a1d2-horizon-secret-key\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.546448 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbnnp\" (UniqueName: \"kubernetes.io/projected/47794116-db67-4451-858c-bb6fea64a21c-kube-api-access-lbnnp\") pod \"placement-db-sync-s7fh4\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.549159 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt4mk\" (UniqueName: \"kubernetes.io/projected/42416335-c299-401c-b732-ae75dfa6a1d2-kube-api-access-wt4mk\") pod \"horizon-79c9bb54c7-hz8j2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.559085 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ttghn"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.561983 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhj94\" (UniqueName: \"kubernetes.io/projected/4cdde67a-2aad-4ef4-ae1b-db394a522f32-kube-api-access-hhj94\") pod \"dnsmasq-dns-7f57676dc-brk7t\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.586956 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.587846 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.590853 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-db-sync-config-data\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.590934 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-combined-ca-bundle\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.590995 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ntfm\" (UniqueName: \"kubernetes.io/projected/0e670ed1-808d-4d6c-aa73-106476663b46-kube-api-access-5ntfm\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.624436 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.632599 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.672820 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7fh4" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.692563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-db-sync-config-data\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.692688 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-combined-ca-bundle\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.692769 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ntfm\" (UniqueName: \"kubernetes.io/projected/0e670ed1-808d-4d6c-aa73-106476663b46-kube-api-access-5ntfm\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.699016 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-db-sync-config-data\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.700244 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-combined-ca-bundle\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.712503 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ntfm\" (UniqueName: \"kubernetes.io/projected/0e670ed1-808d-4d6c-aa73-106476663b46-kube-api-access-5ntfm\") pod \"barbican-db-sync-ttghn\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.739447 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.784546 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.801643 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ttghn" Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.894328 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rsfkt"] Oct 08 15:33:57 crc kubenswrapper[4945]: I1008 15:33:57.905156 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56bd54c5ff-jwcfl"] Oct 08 15:33:57 crc kubenswrapper[4945]: W1008 15:33:57.939830 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab99afa3_52cb_46d4_9a6e_5e00b5a290d5.slice/crio-a329b86eed17cee73f22807e8bf2d0a390519762da28d614bcaaf3f0cc6def5c WatchSource:0}: Error finding container a329b86eed17cee73f22807e8bf2d0a390519762da28d614bcaaf3f0cc6def5c: Status 404 returned error can't find the container with id a329b86eed17cee73f22807e8bf2d0a390519762da28d614bcaaf3f0cc6def5c Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.175038 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68559d9d9f-k46pc"] Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.264344 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.327274 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ztjf5"] Oct 08 15:33:58 crc kubenswrapper[4945]: W1008 15:33:58.344796 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dec4ad0_549d_49a0_9749_86b21bbaf987.slice/crio-c3d7f8797e872ee69b65eef0aeb9ba409af0ff507f794bfc3e812bbeb81bcc8b WatchSource:0}: Error finding container c3d7f8797e872ee69b65eef0aeb9ba409af0ff507f794bfc3e812bbeb81bcc8b: Status 404 returned error can't find the container with id c3d7f8797e872ee69b65eef0aeb9ba409af0ff507f794bfc3e812bbeb81bcc8b Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.347672 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s7fh4"] Oct 08 15:33:58 crc kubenswrapper[4945]: W1008 15:33:58.353750 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47794116_db67_4451_858c_bb6fea64a21c.slice/crio-b854e7f15557e6a4631c07fa185cbdcf9c7ee4742b883fd5d7e0a5c89f5e0ada WatchSource:0}: Error finding container b854e7f15557e6a4631c07fa185cbdcf9c7ee4742b883fd5d7e0a5c89f5e0ada: Status 404 returned error can't find the container with id b854e7f15557e6a4631c07fa185cbdcf9c7ee4742b883fd5d7e0a5c89f5e0ada Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.412144 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pfbzj"] Oct 08 15:33:58 crc kubenswrapper[4945]: W1008 15:33:58.433565 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1be99142_5236_4365_8a45_70847b54b883.slice/crio-22fae66bba0892d3555bea57b50948363a074bc1b0dcb692063a6c5f2b8446bd WatchSource:0}: Error finding container 22fae66bba0892d3555bea57b50948363a074bc1b0dcb692063a6c5f2b8446bd: Status 404 returned error can't find the container with id 22fae66bba0892d3555bea57b50948363a074bc1b0dcb692063a6c5f2b8446bd Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.559626 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f57676dc-brk7t"] Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.562509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rsfkt" event={"ID":"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6","Type":"ContainerStarted","Data":"9481b142aff7c535dd1312fc6e040c7ce6eb4d1b07e29e868578ce368cd06064"} Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.562558 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rsfkt" event={"ID":"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6","Type":"ContainerStarted","Data":"bf91391dc4a250d47e8889b40726c2dd35d6d75d493348b2eea6da9d804814ee"} Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.564521 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7fh4" event={"ID":"47794116-db67-4451-858c-bb6fea64a21c","Type":"ContainerStarted","Data":"b854e7f15557e6a4631c07fa185cbdcf9c7ee4742b883fd5d7e0a5c89f5e0ada"} Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.568639 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68559d9d9f-k46pc" event={"ID":"6a3b7ec6-9364-4d86-981e-e882eb997e41","Type":"ContainerStarted","Data":"b66472c65068950584409e56afdfc5857bab6defe03615a6d62b56300e970cea"} Oct 08 15:33:58 crc kubenswrapper[4945]: W1008 15:33:58.572808 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cdde67a_2aad_4ef4_ae1b_db394a522f32.slice/crio-b56415ed8bc5ff59a2bc104085cfeb659fd09b8c11a8d36d28d9190e0d1011f9 WatchSource:0}: Error finding container b56415ed8bc5ff59a2bc104085cfeb659fd09b8c11a8d36d28d9190e0d1011f9: Status 404 returned error can't find the container with id b56415ed8bc5ff59a2bc104085cfeb659fd09b8c11a8d36d28d9190e0d1011f9 Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.573743 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerStarted","Data":"730a0ad4c6a64576b5f645e20ce116a280249b2edd000429172391f8f5357274"} Oct 08 15:33:58 crc kubenswrapper[4945]: W1008 15:33:58.575046 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e670ed1_808d_4d6c_aa73_106476663b46.slice/crio-ca847cdf5fa40e87a3bdc5754354c37bace6da9a51dfba4244a7dd53edd4e0bb WatchSource:0}: Error finding container ca847cdf5fa40e87a3bdc5754354c37bace6da9a51dfba4244a7dd53edd4e0bb: Status 404 returned error can't find the container with id ca847cdf5fa40e87a3bdc5754354c37bace6da9a51dfba4244a7dd53edd4e0bb Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.594793 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pfbzj" event={"ID":"1be99142-5236-4365-8a45-70847b54b883","Type":"ContainerStarted","Data":"22fae66bba0892d3555bea57b50948363a074bc1b0dcb692063a6c5f2b8446bd"} Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.597225 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ztjf5" event={"ID":"3dec4ad0-549d-49a0-9749-86b21bbaf987","Type":"ContainerStarted","Data":"c3d7f8797e872ee69b65eef0aeb9ba409af0ff507f794bfc3e812bbeb81bcc8b"} Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.598668 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ttghn"] Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.600415 4945 generic.go:334] "Generic (PLEG): container finished" podID="ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" containerID="66240c3b9683c4ba58edd285e03a664a4d9547c3e094ae7443a79ad8f3a4b096" exitCode=0 Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.600448 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" event={"ID":"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5","Type":"ContainerDied","Data":"66240c3b9683c4ba58edd285e03a664a4d9547c3e094ae7443a79ad8f3a4b096"} Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.600471 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" event={"ID":"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5","Type":"ContainerStarted","Data":"a329b86eed17cee73f22807e8bf2d0a390519762da28d614bcaaf3f0cc6def5c"} Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.619645 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rsfkt" podStartSLOduration=2.619625988 podStartE2EDuration="2.619625988s" podCreationTimestamp="2025-10-08 15:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:33:58.578147365 +0000 UTC m=+1247.932062266" watchObservedRunningTime="2025-10-08 15:33:58.619625988 +0000 UTC m=+1247.973540889" Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.693765 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79c9bb54c7-hz8j2"] Oct 08 15:33:58 crc kubenswrapper[4945]: W1008 15:33:58.702953 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42416335_c299_401c_b732_ae75dfa6a1d2.slice/crio-7d50b9456b176338e8d4d1a83ea9875bc53341d1908741072246affb60cc90b9 WatchSource:0}: Error finding container 7d50b9456b176338e8d4d1a83ea9875bc53341d1908741072246affb60cc90b9: Status 404 returned error can't find the container with id 7d50b9456b176338e8d4d1a83ea9875bc53341d1908741072246affb60cc90b9 Oct 08 15:33:58 crc kubenswrapper[4945]: I1008 15:33:58.923428 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.025807 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-config\") pod \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.026075 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h472z\" (UniqueName: \"kubernetes.io/projected/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-kube-api-access-h472z\") pod \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.026172 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-nb\") pod \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.026224 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-sb\") pod \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.026249 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-svc\") pod \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.026265 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-swift-storage-0\") pod \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\" (UID: \"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5\") " Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.067529 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" (UID: "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.068283 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-kube-api-access-h472z" (OuterVolumeSpecName: "kube-api-access-h472z") pod "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" (UID: "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5"). InnerVolumeSpecName "kube-api-access-h472z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.068964 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" (UID: "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.083388 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" (UID: "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.086722 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-config" (OuterVolumeSpecName: "config") pod "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" (UID: "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.092926 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" (UID: "ab99afa3-52cb-46d4-9a6e-5e00b5a290d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.143500 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.143535 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h472z\" (UniqueName: \"kubernetes.io/projected/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-kube-api-access-h472z\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.143548 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.143558 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.143570 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.143580 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.613414 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c9bb54c7-hz8j2" event={"ID":"42416335-c299-401c-b732-ae75dfa6a1d2","Type":"ContainerStarted","Data":"7d50b9456b176338e8d4d1a83ea9875bc53341d1908741072246affb60cc90b9"} Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.615845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ztjf5" event={"ID":"3dec4ad0-549d-49a0-9749-86b21bbaf987","Type":"ContainerStarted","Data":"481ea462b2ffe45c8b6a5bf15ad10dc23616f5609ecafac7d281205ea1afd821"} Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.624904 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ttghn" event={"ID":"0e670ed1-808d-4d6c-aa73-106476663b46","Type":"ContainerStarted","Data":"ca847cdf5fa40e87a3bdc5754354c37bace6da9a51dfba4244a7dd53edd4e0bb"} Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.630547 4945 generic.go:334] "Generic (PLEG): container finished" podID="cf47e507-de36-49a3-9428-7104713b4d54" containerID="b027c7df84247536b2b76ba5946c3242f6ae1576ffde04680aeaeff006508020" exitCode=0 Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.630637 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-t8trl" event={"ID":"cf47e507-de36-49a3-9428-7104713b4d54","Type":"ContainerDied","Data":"b027c7df84247536b2b76ba5946c3242f6ae1576ffde04680aeaeff006508020"} Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.631389 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-ztjf5" podStartSLOduration=3.6313688920000002 podStartE2EDuration="3.631368892s" podCreationTimestamp="2025-10-08 15:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:33:59.629728071 +0000 UTC m=+1248.983642972" watchObservedRunningTime="2025-10-08 15:33:59.631368892 +0000 UTC m=+1248.985283793" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.633269 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.633263 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56bd54c5ff-jwcfl" event={"ID":"ab99afa3-52cb-46d4-9a6e-5e00b5a290d5","Type":"ContainerDied","Data":"a329b86eed17cee73f22807e8bf2d0a390519762da28d614bcaaf3f0cc6def5c"} Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.633395 4945 scope.go:117] "RemoveContainer" containerID="66240c3b9683c4ba58edd285e03a664a4d9547c3e094ae7443a79ad8f3a4b096" Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.637044 4945 generic.go:334] "Generic (PLEG): container finished" podID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerID="76b9f3d811279210459af5f4fe10da8845675fa8f383ef3db696a2101a0b5484" exitCode=0 Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.638000 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" event={"ID":"4cdde67a-2aad-4ef4-ae1b-db394a522f32","Type":"ContainerDied","Data":"76b9f3d811279210459af5f4fe10da8845675fa8f383ef3db696a2101a0b5484"} Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.638062 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" event={"ID":"4cdde67a-2aad-4ef4-ae1b-db394a522f32","Type":"ContainerStarted","Data":"b56415ed8bc5ff59a2bc104085cfeb659fd09b8c11a8d36d28d9190e0d1011f9"} Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.815963 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56bd54c5ff-jwcfl"] Oct 08 15:33:59 crc kubenswrapper[4945]: I1008 15:33:59.863807 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56bd54c5ff-jwcfl"] Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.008514 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68559d9d9f-k46pc"] Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.078902 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" path="/var/lib/kubelet/pods/ab99afa3-52cb-46d4-9a6e-5e00b5a290d5/volumes" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.080498 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.080530 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8b84b4ccc-8lnxg"] Oct 08 15:34:00 crc kubenswrapper[4945]: E1008 15:34:00.080876 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" containerName="init" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.080888 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" containerName="init" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.081529 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab99afa3-52cb-46d4-9a6e-5e00b5a290d5" containerName="init" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.082528 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.094448 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8b84b4ccc-8lnxg"] Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.287806 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-horizon-secret-key\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.287887 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-logs\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.287942 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-config-data\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.288013 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zvn6\" (UniqueName: \"kubernetes.io/projected/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-kube-api-access-7zvn6\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.288048 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-scripts\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.391716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-config-data\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.391795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zvn6\" (UniqueName: \"kubernetes.io/projected/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-kube-api-access-7zvn6\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.391826 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-scripts\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.391903 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-horizon-secret-key\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.391917 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-logs\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.392379 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-logs\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.395769 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-config-data\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.398782 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-scripts\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.404913 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-horizon-secret-key\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.411250 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zvn6\" (UniqueName: \"kubernetes.io/projected/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-kube-api-access-7zvn6\") pod \"horizon-8b84b4ccc-8lnxg\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.437916 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.658229 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" event={"ID":"4cdde67a-2aad-4ef4-ae1b-db394a522f32","Type":"ContainerStarted","Data":"0d7f328d4908a76cb92f7f05e073bbad433dd5de97057a23c56a54975d3a98ef"} Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.679829 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" podStartSLOduration=3.67980979 podStartE2EDuration="3.67980979s" podCreationTimestamp="2025-10-08 15:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:34:00.677710606 +0000 UTC m=+1250.031625507" watchObservedRunningTime="2025-10-08 15:34:00.67980979 +0000 UTC m=+1250.033724691" Oct 08 15:34:00 crc kubenswrapper[4945]: I1008 15:34:00.924256 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8b84b4ccc-8lnxg"] Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.107614 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-t8trl" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.220400 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-config-data\") pod \"cf47e507-de36-49a3-9428-7104713b4d54\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.220546 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-db-sync-config-data\") pod \"cf47e507-de36-49a3-9428-7104713b4d54\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.220710 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx6vj\" (UniqueName: \"kubernetes.io/projected/cf47e507-de36-49a3-9428-7104713b4d54-kube-api-access-bx6vj\") pod \"cf47e507-de36-49a3-9428-7104713b4d54\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.220828 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-combined-ca-bundle\") pod \"cf47e507-de36-49a3-9428-7104713b4d54\" (UID: \"cf47e507-de36-49a3-9428-7104713b4d54\") " Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.229488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cf47e507-de36-49a3-9428-7104713b4d54" (UID: "cf47e507-de36-49a3-9428-7104713b4d54"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.257342 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf47e507-de36-49a3-9428-7104713b4d54-kube-api-access-bx6vj" (OuterVolumeSpecName: "kube-api-access-bx6vj") pod "cf47e507-de36-49a3-9428-7104713b4d54" (UID: "cf47e507-de36-49a3-9428-7104713b4d54"). InnerVolumeSpecName "kube-api-access-bx6vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.286725 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf47e507-de36-49a3-9428-7104713b4d54" (UID: "cf47e507-de36-49a3-9428-7104713b4d54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.308949 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-config-data" (OuterVolumeSpecName: "config-data") pod "cf47e507-de36-49a3-9428-7104713b4d54" (UID: "cf47e507-de36-49a3-9428-7104713b4d54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.325713 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx6vj\" (UniqueName: \"kubernetes.io/projected/cf47e507-de36-49a3-9428-7104713b4d54-kube-api-access-bx6vj\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.325740 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.325775 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.325784 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf47e507-de36-49a3-9428-7104713b4d54-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.730452 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8b84b4ccc-8lnxg" event={"ID":"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1","Type":"ContainerStarted","Data":"abca8458077ae29434a7f22f92b8628fe86dea9cb9962650b843a0688d2f5e30"} Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.742592 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-t8trl" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.743224 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-t8trl" event={"ID":"cf47e507-de36-49a3-9428-7104713b4d54","Type":"ContainerDied","Data":"94b44fd5d6163d59aa5a81bbd4f9ce57ac70f714549c8003e98b01aab714588a"} Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.743253 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94b44fd5d6163d59aa5a81bbd4f9ce57ac70f714549c8003e98b01aab714588a" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.743268 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.862043 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:34:01 crc kubenswrapper[4945]: E1008 15:34:01.862465 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf47e507-de36-49a3-9428-7104713b4d54" containerName="watcher-db-sync" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.862485 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf47e507-de36-49a3-9428-7104713b4d54" containerName="watcher-db-sync" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.862679 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf47e507-de36-49a3-9428-7104713b4d54" containerName="watcher-db-sync" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.865657 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.868298 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.868557 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-gtdt6" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.877386 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.928819 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.930251 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.936894 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 08 15:34:01 crc kubenswrapper[4945]: I1008 15:34:01.955676 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.020456 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.030463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.043315 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-logs\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049097 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049151 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlzgl\" (UniqueName: \"kubernetes.io/projected/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-kube-api-access-wlzgl\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049176 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-config-data\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049198 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-logs\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049294 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049487 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-config-data\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049517 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049563 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.049584 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vdq2\" (UniqueName: \"kubernetes.io/projected/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-kube-api-access-9vdq2\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.059592 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.062056 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151068 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnszh\" (UniqueName: \"kubernetes.io/projected/099759e8-5f3f-463b-9be1-5c6ee209ce0a-kube-api-access-xnszh\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151154 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-logs\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/099759e8-5f3f-463b-9be1-5c6ee209ce0a-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151391 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/099759e8-5f3f-463b-9be1-5c6ee209ce0a-config-data\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151453 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151475 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlzgl\" (UniqueName: \"kubernetes.io/projected/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-kube-api-access-wlzgl\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151537 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-config-data\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151564 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-logs\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151572 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-logs\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.151779 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.152074 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-config-data\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.152140 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.152186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.152242 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vdq2\" (UniqueName: \"kubernetes.io/projected/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-kube-api-access-9vdq2\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.152807 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-logs\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.152965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099759e8-5f3f-463b-9be1-5c6ee209ce0a-logs\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.157702 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-config-data\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.160036 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.162738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.163452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.173658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vdq2\" (UniqueName: \"kubernetes.io/projected/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-kube-api-access-9vdq2\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.174315 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.178863 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlzgl\" (UniqueName: \"kubernetes.io/projected/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-kube-api-access-wlzgl\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.182978 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-config-data\") pod \"watcher-api-0\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.192361 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.254722 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/099759e8-5f3f-463b-9be1-5c6ee209ce0a-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.255013 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/099759e8-5f3f-463b-9be1-5c6ee209ce0a-config-data\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.255172 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099759e8-5f3f-463b-9be1-5c6ee209ce0a-logs\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.255194 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnszh\" (UniqueName: \"kubernetes.io/projected/099759e8-5f3f-463b-9be1-5c6ee209ce0a-kube-api-access-xnszh\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.255603 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099759e8-5f3f-463b-9be1-5c6ee209ce0a-logs\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.258873 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/099759e8-5f3f-463b-9be1-5c6ee209ce0a-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.262855 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/099759e8-5f3f-463b-9be1-5c6ee209ce0a-config-data\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.270368 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnszh\" (UniqueName: \"kubernetes.io/projected/099759e8-5f3f-463b-9be1-5c6ee209ce0a-kube-api-access-xnszh\") pod \"watcher-applier-0\" (UID: \"099759e8-5f3f-463b-9be1-5c6ee209ce0a\") " pod="openstack/watcher-applier-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.274402 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 15:34:02 crc kubenswrapper[4945]: I1008 15:34:02.378939 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 08 15:34:03 crc kubenswrapper[4945]: I1008 15:34:03.361061 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:34:03 crc kubenswrapper[4945]: I1008 15:34:03.529633 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 08 15:34:03 crc kubenswrapper[4945]: I1008 15:34:03.540829 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:03 crc kubenswrapper[4945]: W1008 15:34:03.608828 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099759e8_5f3f_463b_9be1_5c6ee209ce0a.slice/crio-cbf014fe0c17ed2e80a9583f399907431b4ee60b3fada0ae2da7ee7d782a4e51 WatchSource:0}: Error finding container cbf014fe0c17ed2e80a9583f399907431b4ee60b3fada0ae2da7ee7d782a4e51: Status 404 returned error can't find the container with id cbf014fe0c17ed2e80a9583f399907431b4ee60b3fada0ae2da7ee7d782a4e51 Oct 08 15:34:03 crc kubenswrapper[4945]: I1008 15:34:03.768797 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55","Type":"ContainerStarted","Data":"381123a4aedec215745f9b626e42061ef89280f7d8365266d3122165ece229e4"} Oct 08 15:34:03 crc kubenswrapper[4945]: I1008 15:34:03.771046 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"099759e8-5f3f-463b-9be1-5c6ee209ce0a","Type":"ContainerStarted","Data":"cbf014fe0c17ed2e80a9583f399907431b4ee60b3fada0ae2da7ee7d782a4e51"} Oct 08 15:34:03 crc kubenswrapper[4945]: I1008 15:34:03.772248 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerStarted","Data":"1bed85dc545d31d674ad29548530b2ee65272749454136545d2ff05e79b3ee6f"} Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.533327 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79c9bb54c7-hz8j2"] Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.543554 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58798f7484-jkqnb"] Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.544971 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.546997 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.558974 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58798f7484-jkqnb"] Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.614947 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8b84b4ccc-8lnxg"] Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.630646 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c58b7c96-vjh9q"] Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.632182 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.643646 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c58b7c96-vjh9q"] Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.742993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-combined-ca-bundle\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743056 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-config-data\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-secret-key\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743130 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9fcp\" (UniqueName: \"kubernetes.io/projected/37d6d391-e363-42db-8a0e-4e837e4af4db-kube-api-access-l9fcp\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743164 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-horizon-secret-key\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743201 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37d6d391-e363-42db-8a0e-4e837e4af4db-config-data\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743279 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d6d391-e363-42db-8a0e-4e837e4af4db-logs\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743362 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm2ft\" (UniqueName: \"kubernetes.io/projected/4e99e039-6910-4bca-b0ff-815eba456ebd-kube-api-access-dm2ft\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743447 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37d6d391-e363-42db-8a0e-4e837e4af4db-scripts\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743495 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-tls-certs\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743527 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-horizon-tls-certs\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743555 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-scripts\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743580 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e99e039-6910-4bca-b0ff-815eba456ebd-logs\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.743618 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-combined-ca-bundle\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.811394 4945 generic.go:334] "Generic (PLEG): container finished" podID="3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" containerID="9481b142aff7c535dd1312fc6e040c7ce6eb4d1b07e29e868578ce368cd06064" exitCode=0 Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.811433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rsfkt" event={"ID":"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6","Type":"ContainerDied","Data":"9481b142aff7c535dd1312fc6e040c7ce6eb4d1b07e29e868578ce368cd06064"} Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.845269 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-combined-ca-bundle\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.845328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-combined-ca-bundle\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.845404 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-config-data\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-secret-key\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9fcp\" (UniqueName: \"kubernetes.io/projected/37d6d391-e363-42db-8a0e-4e837e4af4db-kube-api-access-l9fcp\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-horizon-secret-key\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846246 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37d6d391-e363-42db-8a0e-4e837e4af4db-config-data\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846275 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d6d391-e363-42db-8a0e-4e837e4af4db-logs\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846305 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm2ft\" (UniqueName: \"kubernetes.io/projected/4e99e039-6910-4bca-b0ff-815eba456ebd-kube-api-access-dm2ft\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846345 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37d6d391-e363-42db-8a0e-4e837e4af4db-scripts\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846370 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-tls-certs\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846391 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-horizon-tls-certs\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846409 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-scripts\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846426 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e99e039-6910-4bca-b0ff-815eba456ebd-logs\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846710 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-config-data\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e99e039-6910-4bca-b0ff-815eba456ebd-logs\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.846987 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d6d391-e363-42db-8a0e-4e837e4af4db-logs\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.848622 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37d6d391-e363-42db-8a0e-4e837e4af4db-scripts\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.849217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-scripts\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.850288 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37d6d391-e363-42db-8a0e-4e837e4af4db-config-data\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.855965 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-horizon-secret-key\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.857342 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-horizon-tls-certs\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.857466 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-tls-certs\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.859329 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-combined-ca-bundle\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.859716 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d6d391-e363-42db-8a0e-4e837e4af4db-combined-ca-bundle\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.873675 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm2ft\" (UniqueName: \"kubernetes.io/projected/4e99e039-6910-4bca-b0ff-815eba456ebd-kube-api-access-dm2ft\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.880088 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-secret-key\") pod \"horizon-58798f7484-jkqnb\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.889649 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9fcp\" (UniqueName: \"kubernetes.io/projected/37d6d391-e363-42db-8a0e-4e837e4af4db-kube-api-access-l9fcp\") pod \"horizon-7c58b7c96-vjh9q\" (UID: \"37d6d391-e363-42db-8a0e-4e837e4af4db\") " pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:06 crc kubenswrapper[4945]: I1008 15:34:06.952979 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:07 crc kubenswrapper[4945]: I1008 15:34:07.161884 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:07 crc kubenswrapper[4945]: I1008 15:34:07.742571 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:34:07 crc kubenswrapper[4945]: I1008 15:34:07.825316 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9c464d5c-8j67t"] Oct 08 15:34:07 crc kubenswrapper[4945]: I1008 15:34:07.825649 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" containerID="cri-o://31f9c9a3ea6dcbccb34ddd8909963c1f4777115961eb032e19e6244e53314035" gracePeriod=10 Oct 08 15:34:08 crc kubenswrapper[4945]: I1008 15:34:08.836199 4945 generic.go:334] "Generic (PLEG): container finished" podID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerID="31f9c9a3ea6dcbccb34ddd8909963c1f4777115961eb032e19e6244e53314035" exitCode=0 Oct 08 15:34:08 crc kubenswrapper[4945]: I1008 15:34:08.836252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" event={"ID":"ade81fed-c2cb-46a1-9623-96bcc42b65b7","Type":"ContainerDied","Data":"31f9c9a3ea6dcbccb34ddd8909963c1f4777115961eb032e19e6244e53314035"} Oct 08 15:34:12 crc kubenswrapper[4945]: I1008 15:34:12.430147 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.319494 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.430287 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.448900 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-combined-ca-bundle\") pod \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.448986 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-fernet-keys\") pod \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.449036 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-credential-keys\") pod \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.449096 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-config-data\") pod \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.449179 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-scripts\") pod \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.449320 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz5kc\" (UniqueName: \"kubernetes.io/projected/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-kube-api-access-tz5kc\") pod \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\" (UID: \"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6\") " Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.469785 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" (UID: "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.469878 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-kube-api-access-tz5kc" (OuterVolumeSpecName: "kube-api-access-tz5kc") pod "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" (UID: "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6"). InnerVolumeSpecName "kube-api-access-tz5kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.471457 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-scripts" (OuterVolumeSpecName: "scripts") pod "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" (UID: "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.471891 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" (UID: "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.485685 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-config-data" (OuterVolumeSpecName: "config-data") pod "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" (UID: "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.494037 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" (UID: "3f7ef194-3d84-4dd4-84f9-fb7a5012fac6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.541418 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.541892 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.542044 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54h5b4hb8hbh57bhbdh77h55dhc4h685h599hfdhf8h654h699h5bch658h564h57fh684h5d7h76h5dbhbdh575h664hcdh564h568h5cch95h75q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qfz87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-68559d9d9f-k46pc_openstack(6a3b7ec6-9364-4d86-981e-e882eb997e41): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.544236 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-68559d9d9f-k46pc" podUID="6a3b7ec6-9364-4d86-981e-e882eb997e41" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.551178 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.551219 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz5kc\" (UniqueName: \"kubernetes.io/projected/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-kube-api-access-tz5kc\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.551235 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.551246 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.551257 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.551267 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.573469 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.573526 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.573655 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n6bh557h698h56h599h84h669h68hddhcfh694h665h94h6h645h548h68fh68bh64dhf5h77hffh55h65ch574h59fh5b5h54bh8hfbh8ch64cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wt4mk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-79c9bb54c7-hz8j2_openstack(42416335-c299-401c-b732-ae75dfa6a1d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.585684 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-79c9bb54c7-hz8j2" podUID="42416335-c299-401c-b732-ae75dfa6a1d2" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.677829 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.677897 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.678039 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbchfch68bh697hbfhf7h548h59bhd6hc5h55h58h69h5fch656h64bh59dh674h58fh5bdh7fh5fchc5h645h54dh5c8hc6h676h598h77h54dh58dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7zvn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-8b84b4ccc-8lnxg_openstack(26ca94e7-5c48-41d0-8c37-f0748ee7f6f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.680851 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-8b84b4ccc-8lnxg" podUID="26ca94e7-5c48-41d0-8c37-f0748ee7f6f1" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.813844 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-watcher-applier:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.813896 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-watcher-applier:watcher_latest" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.814311 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:watcher-applier,Image:38.102.83.163:5001/podified-master-centos10/openstack-watcher-applier:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5hf6h65h5c8h676h545h55bh599h596h56h9bh687h596h646h5b9h5fh685h67dh77h5cbh54dh57h5d4h57dh665hdfh5fhfh695h9dh689h688q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:watcher-applier-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/watcher,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xnszh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pgrep -r DRST watcher-applier],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pgrep -r DRST watcher-applier],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42451,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pgrep -r DRST watcher-applier],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-applier-0_openstack(099759e8-5f3f-463b-9be1-5c6ee209ce0a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.815522 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-applier\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/watcher-applier-0" podUID="099759e8-5f3f-463b-9be1-5c6ee209ce0a" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.945167 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rsfkt" Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.945346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rsfkt" event={"ID":"3f7ef194-3d84-4dd4-84f9-fb7a5012fac6","Type":"ContainerDied","Data":"bf91391dc4a250d47e8889b40726c2dd35d6d75d493348b2eea6da9d804814ee"} Oct 08 15:34:17 crc kubenswrapper[4945]: I1008 15:34:17.945876 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf91391dc4a250d47e8889b40726c2dd35d6d75d493348b2eea6da9d804814ee" Oct 08 15:34:17 crc kubenswrapper[4945]: E1008 15:34:17.948852 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-applier\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-watcher-applier:watcher_latest\\\"\"" pod="openstack/watcher-applier-0" podUID="099759e8-5f3f-463b-9be1-5c6ee209ce0a" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.398664 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rsfkt"] Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.404742 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rsfkt"] Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.518747 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-qkgj2"] Oct 08 15:34:18 crc kubenswrapper[4945]: E1008 15:34:18.523547 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" containerName="keystone-bootstrap" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.523583 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" containerName="keystone-bootstrap" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.523927 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" containerName="keystone-bootstrap" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.524724 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.531139 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qkgj2"] Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.540598 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.540831 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.555008 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.564392 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7rf" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.583029 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-fernet-keys\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.583081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-combined-ca-bundle\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.583110 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-credential-keys\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.583179 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-config-data\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.583220 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv9ds\" (UniqueName: \"kubernetes.io/projected/30020f4e-d113-4b93-baa0-553942be7a4d-kube-api-access-hv9ds\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.583243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-scripts\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.684393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-config-data\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.684450 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv9ds\" (UniqueName: \"kubernetes.io/projected/30020f4e-d113-4b93-baa0-553942be7a4d-kube-api-access-hv9ds\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.684478 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-scripts\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.684525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-fernet-keys\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.684555 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-combined-ca-bundle\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.684575 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-credential-keys\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.696960 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-scripts\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.697314 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-combined-ca-bundle\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.697564 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-credential-keys\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.698768 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-fernet-keys\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.699732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-config-data\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.722581 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv9ds\" (UniqueName: \"kubernetes.io/projected/30020f4e-d113-4b93-baa0-553942be7a4d-kube-api-access-hv9ds\") pod \"keystone-bootstrap-qkgj2\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.854591 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:18 crc kubenswrapper[4945]: I1008 15:34:18.956593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55","Type":"ContainerStarted","Data":"0cce6c7dc8a496d9735089da371c1f6d51223aae0fa2bf7b46a8ecf950fd2abc"} Oct 08 15:34:20 crc kubenswrapper[4945]: I1008 15:34:20.036312 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f7ef194-3d84-4dd4-84f9-fb7a5012fac6" path="/var/lib/kubelet/pods/3f7ef194-3d84-4dd4-84f9-fb7a5012fac6/volumes" Oct 08 15:34:27 crc kubenswrapper[4945]: I1008 15:34:27.430695 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Oct 08 15:34:27 crc kubenswrapper[4945]: I1008 15:34:27.431267 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:34:30 crc kubenswrapper[4945]: E1008 15:34:30.510408 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 08 15:34:30 crc kubenswrapper[4945]: E1008 15:34:30.511074 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 08 15:34:30 crc kubenswrapper[4945]: E1008 15:34:30.511300 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.163:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pclt8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-pfbzj_openstack(1be99142-5236-4365-8a45-70847b54b883): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:30 crc kubenswrapper[4945]: E1008 15:34:30.512585 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-pfbzj" podUID="1be99142-5236-4365-8a45-70847b54b883" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.073974 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-pfbzj" podUID="1be99142-5236-4365-8a45-70847b54b883" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.669281 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.669360 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.669537 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qr722,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-4bnbb_openstack(0646f7b0-6abf-436c-976f-59ab119525ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.670727 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-4bnbb" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.720949 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-placement-api:watcher_latest" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.721465 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-placement-api:watcher_latest" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.721669 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:38.102.83.163:5001/podified-master-centos10/openstack-placement-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-s7fh4_openstack(47794116-db67-4451-858c-bb6fea64a21c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:31 crc kubenswrapper[4945]: E1008 15:34:31.722937 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-s7fh4" podUID="47794116-db67-4451-858c-bb6fea64a21c" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.814933 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.822749 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.831761 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.842348 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855318 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-scripts\") pod \"42416335-c299-401c-b732-ae75dfa6a1d2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855411 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-svc\") pod \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855507 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zvn6\" (UniqueName: \"kubernetes.io/projected/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-kube-api-access-7zvn6\") pod \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855575 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-config-data\") pod \"6a3b7ec6-9364-4d86-981e-e882eb997e41\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855610 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-config\") pod \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855653 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-scripts\") pod \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855701 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-logs\") pod \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855741 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfz87\" (UniqueName: \"kubernetes.io/projected/6a3b7ec6-9364-4d86-981e-e882eb997e41-kube-api-access-qfz87\") pod \"6a3b7ec6-9364-4d86-981e-e882eb997e41\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855797 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fncz6\" (UniqueName: \"kubernetes.io/projected/ade81fed-c2cb-46a1-9623-96bcc42b65b7-kube-api-access-fncz6\") pod \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855854 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-nb\") pod \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855884 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-horizon-secret-key\") pod \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.855920 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a3b7ec6-9364-4d86-981e-e882eb997e41-horizon-secret-key\") pod \"6a3b7ec6-9364-4d86-981e-e882eb997e41\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856039 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-scripts\") pod \"6a3b7ec6-9364-4d86-981e-e882eb997e41\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856080 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-swift-storage-0\") pod \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856169 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-sb\") pod \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\" (UID: \"ade81fed-c2cb-46a1-9623-96bcc42b65b7\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856202 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-config-data\") pod \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\" (UID: \"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856235 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-config-data\") pod \"42416335-c299-401c-b732-ae75dfa6a1d2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856272 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt4mk\" (UniqueName: \"kubernetes.io/projected/42416335-c299-401c-b732-ae75dfa6a1d2-kube-api-access-wt4mk\") pod \"42416335-c299-401c-b732-ae75dfa6a1d2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856316 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3b7ec6-9364-4d86-981e-e882eb997e41-logs\") pod \"6a3b7ec6-9364-4d86-981e-e882eb997e41\" (UID: \"6a3b7ec6-9364-4d86-981e-e882eb997e41\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856362 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/42416335-c299-401c-b732-ae75dfa6a1d2-horizon-secret-key\") pod \"42416335-c299-401c-b732-ae75dfa6a1d2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.856397 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42416335-c299-401c-b732-ae75dfa6a1d2-logs\") pod \"42416335-c299-401c-b732-ae75dfa6a1d2\" (UID: \"42416335-c299-401c-b732-ae75dfa6a1d2\") " Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.858534 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42416335-c299-401c-b732-ae75dfa6a1d2-logs" (OuterVolumeSpecName: "logs") pod "42416335-c299-401c-b732-ae75dfa6a1d2" (UID: "42416335-c299-401c-b732-ae75dfa6a1d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.858744 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-scripts" (OuterVolumeSpecName: "scripts") pod "42416335-c299-401c-b732-ae75dfa6a1d2" (UID: "42416335-c299-401c-b732-ae75dfa6a1d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.860433 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-logs" (OuterVolumeSpecName: "logs") pod "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1" (UID: "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.860826 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-scripts" (OuterVolumeSpecName: "scripts") pod "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1" (UID: "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.861932 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-config-data" (OuterVolumeSpecName: "config-data") pod "6a3b7ec6-9364-4d86-981e-e882eb997e41" (UID: "6a3b7ec6-9364-4d86-981e-e882eb997e41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.862022 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-config-data" (OuterVolumeSpecName: "config-data") pod "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1" (UID: "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.862597 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-scripts" (OuterVolumeSpecName: "scripts") pod "6a3b7ec6-9364-4d86-981e-e882eb997e41" (UID: "6a3b7ec6-9364-4d86-981e-e882eb997e41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.863086 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-config-data" (OuterVolumeSpecName: "config-data") pod "42416335-c299-401c-b732-ae75dfa6a1d2" (UID: "42416335-c299-401c-b732-ae75dfa6a1d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.865660 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a3b7ec6-9364-4d86-981e-e882eb997e41-logs" (OuterVolumeSpecName: "logs") pod "6a3b7ec6-9364-4d86-981e-e882eb997e41" (UID: "6a3b7ec6-9364-4d86-981e-e882eb997e41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.866433 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-kube-api-access-7zvn6" (OuterVolumeSpecName: "kube-api-access-7zvn6") pod "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1" (UID: "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1"). InnerVolumeSpecName "kube-api-access-7zvn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.874247 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ade81fed-c2cb-46a1-9623-96bcc42b65b7-kube-api-access-fncz6" (OuterVolumeSpecName: "kube-api-access-fncz6") pod "ade81fed-c2cb-46a1-9623-96bcc42b65b7" (UID: "ade81fed-c2cb-46a1-9623-96bcc42b65b7"). InnerVolumeSpecName "kube-api-access-fncz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.878473 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1" (UID: "26ca94e7-5c48-41d0-8c37-f0748ee7f6f1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.879664 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42416335-c299-401c-b732-ae75dfa6a1d2-kube-api-access-wt4mk" (OuterVolumeSpecName: "kube-api-access-wt4mk") pod "42416335-c299-401c-b732-ae75dfa6a1d2" (UID: "42416335-c299-401c-b732-ae75dfa6a1d2"). InnerVolumeSpecName "kube-api-access-wt4mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.881429 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a3b7ec6-9364-4d86-981e-e882eb997e41-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6a3b7ec6-9364-4d86-981e-e882eb997e41" (UID: "6a3b7ec6-9364-4d86-981e-e882eb997e41"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.896397 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42416335-c299-401c-b732-ae75dfa6a1d2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "42416335-c299-401c-b732-ae75dfa6a1d2" (UID: "42416335-c299-401c-b732-ae75dfa6a1d2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.925139 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3b7ec6-9364-4d86-981e-e882eb997e41-kube-api-access-qfz87" (OuterVolumeSpecName: "kube-api-access-qfz87") pod "6a3b7ec6-9364-4d86-981e-e882eb997e41" (UID: "6a3b7ec6-9364-4d86-981e-e882eb997e41"). InnerVolumeSpecName "kube-api-access-qfz87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.952608 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ade81fed-c2cb-46a1-9623-96bcc42b65b7" (UID: "ade81fed-c2cb-46a1-9623-96bcc42b65b7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.954331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ade81fed-c2cb-46a1-9623-96bcc42b65b7" (UID: "ade81fed-c2cb-46a1-9623-96bcc42b65b7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958576 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958612 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958624 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958639 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt4mk\" (UniqueName: \"kubernetes.io/projected/42416335-c299-401c-b732-ae75dfa6a1d2-kube-api-access-wt4mk\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958652 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3b7ec6-9364-4d86-981e-e882eb997e41-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958663 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/42416335-c299-401c-b732-ae75dfa6a1d2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958674 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42416335-c299-401c-b732-ae75dfa6a1d2-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958685 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42416335-c299-401c-b732-ae75dfa6a1d2-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958697 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zvn6\" (UniqueName: \"kubernetes.io/projected/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-kube-api-access-7zvn6\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958707 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958717 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958729 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958740 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfz87\" (UniqueName: \"kubernetes.io/projected/6a3b7ec6-9364-4d86-981e-e882eb997e41-kube-api-access-qfz87\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958751 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fncz6\" (UniqueName: \"kubernetes.io/projected/ade81fed-c2cb-46a1-9623-96bcc42b65b7-kube-api-access-fncz6\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958763 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958774 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a3b7ec6-9364-4d86-981e-e882eb997e41-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958788 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a3b7ec6-9364-4d86-981e-e882eb997e41-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.958799 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.965051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-config" (OuterVolumeSpecName: "config") pod "ade81fed-c2cb-46a1-9623-96bcc42b65b7" (UID: "ade81fed-c2cb-46a1-9623-96bcc42b65b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.967864 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ade81fed-c2cb-46a1-9623-96bcc42b65b7" (UID: "ade81fed-c2cb-46a1-9623-96bcc42b65b7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:31 crc kubenswrapper[4945]: I1008 15:34:31.976156 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ade81fed-c2cb-46a1-9623-96bcc42b65b7" (UID: "ade81fed-c2cb-46a1-9623-96bcc42b65b7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.060327 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.060357 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.060366 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade81fed-c2cb-46a1-9623-96bcc42b65b7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.080120 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68559d9d9f-k46pc" event={"ID":"6a3b7ec6-9364-4d86-981e-e882eb997e41","Type":"ContainerDied","Data":"b66472c65068950584409e56afdfc5857bab6defe03615a6d62b56300e970cea"} Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.080134 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68559d9d9f-k46pc" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.082260 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c9bb54c7-hz8j2" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.082268 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c9bb54c7-hz8j2" event={"ID":"42416335-c299-401c-b732-ae75dfa6a1d2","Type":"ContainerDied","Data":"7d50b9456b176338e8d4d1a83ea9875bc53341d1908741072246affb60cc90b9"} Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.083260 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8b84b4ccc-8lnxg" event={"ID":"26ca94e7-5c48-41d0-8c37-f0748ee7f6f1","Type":"ContainerDied","Data":"abca8458077ae29434a7f22f92b8628fe86dea9cb9962650b843a0688d2f5e30"} Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.083312 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8b84b4ccc-8lnxg" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.085528 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" event={"ID":"ade81fed-c2cb-46a1-9623-96bcc42b65b7","Type":"ContainerDied","Data":"3060ab35805dc8b9be051b8d73ae26ec36e802cca3b4fbdbeaa84d5b6c39542e"} Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.085579 4945 scope.go:117] "RemoveContainer" containerID="31f9c9a3ea6dcbccb34ddd8909963c1f4777115961eb032e19e6244e53314035" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.085588 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.086758 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-placement-api:watcher_latest\\\"\"" pod="openstack/placement-db-sync-s7fh4" podUID="47794116-db67-4451-858c-bb6fea64a21c" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.150894 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8b84b4ccc-8lnxg"] Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.161092 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8b84b4ccc-8lnxg"] Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.180659 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68559d9d9f-k46pc"] Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.194290 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68559d9d9f-k46pc"] Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.214693 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79c9bb54c7-hz8j2"] Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.228163 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79c9bb54c7-hz8j2"] Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.234451 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9c464d5c-8j67t"] Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.241031 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d9c464d5c-8j67t"] Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.358868 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.359272 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.359404 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.163:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5ntfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-ttghn_openstack(0e670ed1-808d-4d6c-aa73-106476663b46): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.360605 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-ttghn" podUID="0e670ed1-808d-4d6c-aa73-106476663b46" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.431340 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d9c464d5c-8j67t" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.613652 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.613709 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 08 15:34:32 crc kubenswrapper[4945]: E1008 15:34:32.613836 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.102.83.163:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb4hf9hfbh5dh5dh66ch59dh669h555hb4h9dh97h98h94h6chbfh7bh687hd4h665h55dhc8h57dh5c6h5bh664h55fh587h576h557hd7h5d7q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmnck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(86db9194-a67b-4491-b0f1-d03bafb3e353): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:34:32 crc kubenswrapper[4945]: I1008 15:34:32.946014 4945 scope.go:117] "RemoveContainer" containerID="811e0a6786dd761db08ddf61a496caaf55bd7adfd28f033b5683c051766e9e55" Oct 08 15:34:33 crc kubenswrapper[4945]: E1008 15:34:33.108201 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-ttghn" podUID="0e670ed1-808d-4d6c-aa73-106476663b46" Oct 08 15:34:33 crc kubenswrapper[4945]: I1008 15:34:33.252978 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58798f7484-jkqnb"] Oct 08 15:34:33 crc kubenswrapper[4945]: I1008 15:34:33.367329 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c58b7c96-vjh9q"] Oct 08 15:34:33 crc kubenswrapper[4945]: I1008 15:34:33.440331 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qkgj2"] Oct 08 15:34:33 crc kubenswrapper[4945]: W1008 15:34:33.539132 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30020f4e_d113_4b93_baa0_553942be7a4d.slice/crio-3c54c598057ff2f288d5c389c16d4fe83a9e5151d1ba6a30d0d7d63aa16b8694 WatchSource:0}: Error finding container 3c54c598057ff2f288d5c389c16d4fe83a9e5151d1ba6a30d0d7d63aa16b8694: Status 404 returned error can't find the container with id 3c54c598057ff2f288d5c389c16d4fe83a9e5151d1ba6a30d0d7d63aa16b8694 Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.033986 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26ca94e7-5c48-41d0-8c37-f0748ee7f6f1" path="/var/lib/kubelet/pods/26ca94e7-5c48-41d0-8c37-f0748ee7f6f1/volumes" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.034774 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42416335-c299-401c-b732-ae75dfa6a1d2" path="/var/lib/kubelet/pods/42416335-c299-401c-b732-ae75dfa6a1d2/volumes" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.035358 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3b7ec6-9364-4d86-981e-e882eb997e41" path="/var/lib/kubelet/pods/6a3b7ec6-9364-4d86-981e-e882eb997e41/volumes" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.035785 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" path="/var/lib/kubelet/pods/ade81fed-c2cb-46a1-9623-96bcc42b65b7/volumes" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.106594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerStarted","Data":"071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.108067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c58b7c96-vjh9q" event={"ID":"37d6d391-e363-42db-8a0e-4e837e4af4db","Type":"ContainerStarted","Data":"8886ddf4335a0f0fd0b3dd57eed02f642d2d50f812ab3285e000fb4945ccf2d1"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.108100 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c58b7c96-vjh9q" event={"ID":"37d6d391-e363-42db-8a0e-4e837e4af4db","Type":"ContainerStarted","Data":"9fee0421ef52559ed159623dcbd686ffd429c985fcce17f5e23c382fa40ae2fa"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.110348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55","Type":"ContainerStarted","Data":"b407ee8c9926a229b916162eea1d5ec04d48149647ab7b7b7974cb5fc1f2441f"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.110982 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.115728 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"099759e8-5f3f-463b-9be1-5c6ee209ce0a","Type":"ContainerStarted","Data":"801ad74434df5a5dd044039877275a86f76f6117f5cbff628b9f4ea6ac8e4a45"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.116968 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerStarted","Data":"5e411c682b3d7d408df995f49aeca0a5f53e009064805c8b6559892f2c9d1713"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.118389 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58798f7484-jkqnb" event={"ID":"4e99e039-6910-4bca-b0ff-815eba456ebd","Type":"ContainerStarted","Data":"39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.118430 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58798f7484-jkqnb" event={"ID":"4e99e039-6910-4bca-b0ff-815eba456ebd","Type":"ContainerStarted","Data":"1ed999df9b090cc5d890c6dce8857588abd68afdb071c73d457198c564b51b0c"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.120038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qkgj2" event={"ID":"30020f4e-d113-4b93-baa0-553942be7a4d","Type":"ContainerStarted","Data":"4f80b380f1351f165e1e0ed9bae4191556d2b142ef136edc04bb5bcfb88ba0ed"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.120067 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qkgj2" event={"ID":"30020f4e-d113-4b93-baa0-553942be7a4d","Type":"ContainerStarted","Data":"3c54c598057ff2f288d5c389c16d4fe83a9e5151d1ba6a30d0d7d63aa16b8694"} Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.138440 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=33.138419661 podStartE2EDuration="33.138419661s" podCreationTimestamp="2025-10-08 15:34:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:34:34.131673091 +0000 UTC m=+1283.485588002" watchObservedRunningTime="2025-10-08 15:34:34.138419661 +0000 UTC m=+1283.492334562" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.158988 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=3.521839879 podStartE2EDuration="33.158972437s" podCreationTimestamp="2025-10-08 15:34:01 +0000 UTC" firstStartedPulling="2025-10-08 15:34:03.386339224 +0000 UTC m=+1252.740254125" lastFinishedPulling="2025-10-08 15:34:33.023471782 +0000 UTC m=+1282.377386683" observedRunningTime="2025-10-08 15:34:34.149344635 +0000 UTC m=+1283.503259536" watchObservedRunningTime="2025-10-08 15:34:34.158972437 +0000 UTC m=+1283.512887338" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.164819 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-qkgj2" podStartSLOduration=16.164804643 podStartE2EDuration="16.164804643s" podCreationTimestamp="2025-10-08 15:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:34:34.164143917 +0000 UTC m=+1283.518058818" watchObservedRunningTime="2025-10-08 15:34:34.164804643 +0000 UTC m=+1283.518719544" Oct 08 15:34:34 crc kubenswrapper[4945]: I1008 15:34:34.184078 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=3.712420659 podStartE2EDuration="33.184060117s" podCreationTimestamp="2025-10-08 15:34:01 +0000 UTC" firstStartedPulling="2025-10-08 15:34:03.61178731 +0000 UTC m=+1252.965702201" lastFinishedPulling="2025-10-08 15:34:33.083426758 +0000 UTC m=+1282.437341659" observedRunningTime="2025-10-08 15:34:34.181411021 +0000 UTC m=+1283.535325922" watchObservedRunningTime="2025-10-08 15:34:34.184060117 +0000 UTC m=+1283.537975018" Oct 08 15:34:35 crc kubenswrapper[4945]: I1008 15:34:35.131384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58798f7484-jkqnb" event={"ID":"4e99e039-6910-4bca-b0ff-815eba456ebd","Type":"ContainerStarted","Data":"1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b"} Oct 08 15:34:35 crc kubenswrapper[4945]: I1008 15:34:35.136873 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c58b7c96-vjh9q" event={"ID":"37d6d391-e363-42db-8a0e-4e837e4af4db","Type":"ContainerStarted","Data":"82e128d31debed13e44f1f4df2629e287ee85f70f42861c93a22691c0ff6aa0e"} Oct 08 15:34:35 crc kubenswrapper[4945]: I1008 15:34:35.158670 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-58798f7484-jkqnb" podStartSLOduration=28.760851213 podStartE2EDuration="29.158648459s" podCreationTimestamp="2025-10-08 15:34:06 +0000 UTC" firstStartedPulling="2025-10-08 15:34:33.253441352 +0000 UTC m=+1282.607356253" lastFinishedPulling="2025-10-08 15:34:33.651238588 +0000 UTC m=+1283.005153499" observedRunningTime="2025-10-08 15:34:35.152212717 +0000 UTC m=+1284.506127638" watchObservedRunningTime="2025-10-08 15:34:35.158648459 +0000 UTC m=+1284.512563360" Oct 08 15:34:35 crc kubenswrapper[4945]: I1008 15:34:35.186463 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c58b7c96-vjh9q" podStartSLOduration=29.072067393 podStartE2EDuration="29.186442007s" podCreationTimestamp="2025-10-08 15:34:06 +0000 UTC" firstStartedPulling="2025-10-08 15:34:33.536268229 +0000 UTC m=+1282.890183130" lastFinishedPulling="2025-10-08 15:34:33.650642833 +0000 UTC m=+1283.004557744" observedRunningTime="2025-10-08 15:34:35.180914838 +0000 UTC m=+1284.534829749" watchObservedRunningTime="2025-10-08 15:34:35.186442007 +0000 UTC m=+1284.540356908" Oct 08 15:34:36 crc kubenswrapper[4945]: I1008 15:34:36.252577 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 15:34:36 crc kubenswrapper[4945]: I1008 15:34:36.954122 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:36 crc kubenswrapper[4945]: I1008 15:34:36.954175 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:37 crc kubenswrapper[4945]: I1008 15:34:37.164235 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:37 crc kubenswrapper[4945]: I1008 15:34:37.164490 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:37 crc kubenswrapper[4945]: I1008 15:34:37.275420 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 15:34:37 crc kubenswrapper[4945]: I1008 15:34:37.379629 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 08 15:34:39 crc kubenswrapper[4945]: I1008 15:34:39.181883 4945 generic.go:334] "Generic (PLEG): container finished" podID="30020f4e-d113-4b93-baa0-553942be7a4d" containerID="4f80b380f1351f165e1e0ed9bae4191556d2b142ef136edc04bb5bcfb88ba0ed" exitCode=0 Oct 08 15:34:39 crc kubenswrapper[4945]: I1008 15:34:39.181980 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qkgj2" event={"ID":"30020f4e-d113-4b93-baa0-553942be7a4d","Type":"ContainerDied","Data":"4f80b380f1351f165e1e0ed9bae4191556d2b142ef136edc04bb5bcfb88ba0ed"} Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.194685 4945 generic.go:334] "Generic (PLEG): container finished" podID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerID="5e411c682b3d7d408df995f49aeca0a5f53e009064805c8b6559892f2c9d1713" exitCode=1 Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.194777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerDied","Data":"5e411c682b3d7d408df995f49aeca0a5f53e009064805c8b6559892f2c9d1713"} Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.195493 4945 scope.go:117] "RemoveContainer" containerID="5e411c682b3d7d408df995f49aeca0a5f53e009064805c8b6559892f2c9d1713" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.550508 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.723308 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-scripts\") pod \"30020f4e-d113-4b93-baa0-553942be7a4d\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.723399 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-config-data\") pod \"30020f4e-d113-4b93-baa0-553942be7a4d\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.723531 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv9ds\" (UniqueName: \"kubernetes.io/projected/30020f4e-d113-4b93-baa0-553942be7a4d-kube-api-access-hv9ds\") pod \"30020f4e-d113-4b93-baa0-553942be7a4d\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.723566 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-combined-ca-bundle\") pod \"30020f4e-d113-4b93-baa0-553942be7a4d\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.723599 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-fernet-keys\") pod \"30020f4e-d113-4b93-baa0-553942be7a4d\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.723639 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-credential-keys\") pod \"30020f4e-d113-4b93-baa0-553942be7a4d\" (UID: \"30020f4e-d113-4b93-baa0-553942be7a4d\") " Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.731406 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "30020f4e-d113-4b93-baa0-553942be7a4d" (UID: "30020f4e-d113-4b93-baa0-553942be7a4d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.731446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "30020f4e-d113-4b93-baa0-553942be7a4d" (UID: "30020f4e-d113-4b93-baa0-553942be7a4d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.732195 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30020f4e-d113-4b93-baa0-553942be7a4d-kube-api-access-hv9ds" (OuterVolumeSpecName: "kube-api-access-hv9ds") pod "30020f4e-d113-4b93-baa0-553942be7a4d" (UID: "30020f4e-d113-4b93-baa0-553942be7a4d"). InnerVolumeSpecName "kube-api-access-hv9ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.734316 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-scripts" (OuterVolumeSpecName: "scripts") pod "30020f4e-d113-4b93-baa0-553942be7a4d" (UID: "30020f4e-d113-4b93-baa0-553942be7a4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.752948 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30020f4e-d113-4b93-baa0-553942be7a4d" (UID: "30020f4e-d113-4b93-baa0-553942be7a4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.759257 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-config-data" (OuterVolumeSpecName: "config-data") pod "30020f4e-d113-4b93-baa0-553942be7a4d" (UID: "30020f4e-d113-4b93-baa0-553942be7a4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.826431 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.826663 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.826686 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv9ds\" (UniqueName: \"kubernetes.io/projected/30020f4e-d113-4b93-baa0-553942be7a4d-kube-api-access-hv9ds\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.826703 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.826721 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:40 crc kubenswrapper[4945]: I1008 15:34:40.826738 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30020f4e-d113-4b93-baa0-553942be7a4d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.206881 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerStarted","Data":"ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32"} Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.210973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerStarted","Data":"c29850611bb832f076ebe4db8820b3e4631318cb4c5ab1a4a0f59924f416f1b1"} Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.212770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qkgj2" event={"ID":"30020f4e-d113-4b93-baa0-553942be7a4d","Type":"ContainerDied","Data":"3c54c598057ff2f288d5c389c16d4fe83a9e5151d1ba6a30d0d7d63aa16b8694"} Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.212797 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c54c598057ff2f288d5c389c16d4fe83a9e5151d1ba6a30d0d7d63aa16b8694" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.212886 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qkgj2" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.310251 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-65b767dfc8-l5jkz"] Oct 08 15:34:41 crc kubenswrapper[4945]: E1008 15:34:41.310781 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="init" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.310803 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="init" Oct 08 15:34:41 crc kubenswrapper[4945]: E1008 15:34:41.311312 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30020f4e-d113-4b93-baa0-553942be7a4d" containerName="keystone-bootstrap" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.311343 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="30020f4e-d113-4b93-baa0-553942be7a4d" containerName="keystone-bootstrap" Oct 08 15:34:41 crc kubenswrapper[4945]: E1008 15:34:41.311368 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.311378 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.311867 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade81fed-c2cb-46a1-9623-96bcc42b65b7" containerName="dnsmasq-dns" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.311897 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="30020f4e-d113-4b93-baa0-553942be7a4d" containerName="keystone-bootstrap" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.315402 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.320493 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.320913 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.321127 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.321368 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.321628 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7rf" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.321827 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.341653 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-65b767dfc8-l5jkz"] Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.444644 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-credential-keys\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.444709 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-config-data\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.444784 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76qp6\" (UniqueName: \"kubernetes.io/projected/f748658a-398b-4b6e-9f27-d2a45e6f6827-kube-api-access-76qp6\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.444804 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-scripts\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.444830 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-internal-tls-certs\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.445360 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-fernet-keys\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.445450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-public-tls-certs\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.445480 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-combined-ca-bundle\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-config-data\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547506 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76qp6\" (UniqueName: \"kubernetes.io/projected/f748658a-398b-4b6e-9f27-d2a45e6f6827-kube-api-access-76qp6\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547533 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-scripts\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547566 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-internal-tls-certs\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547605 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-fernet-keys\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547630 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-public-tls-certs\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547652 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-combined-ca-bundle\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.547717 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-credential-keys\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.554052 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-fernet-keys\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.560665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-scripts\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.562441 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-credential-keys\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.563181 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-public-tls-certs\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.563797 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-combined-ca-bundle\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.564871 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-config-data\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.567628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76qp6\" (UniqueName: \"kubernetes.io/projected/f748658a-398b-4b6e-9f27-d2a45e6f6827-kube-api-access-76qp6\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.567703 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f748658a-398b-4b6e-9f27-d2a45e6f6827-internal-tls-certs\") pod \"keystone-65b767dfc8-l5jkz\" (UID: \"f748658a-398b-4b6e-9f27-d2a45e6f6827\") " pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.656140 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:41 crc kubenswrapper[4945]: I1008 15:34:41.918170 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-65b767dfc8-l5jkz"] Oct 08 15:34:41 crc kubenswrapper[4945]: W1008 15:34:41.946185 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf748658a_398b_4b6e_9f27_d2a45e6f6827.slice/crio-cd22b705caa57910fe8b0cc4a2a4875140ee9107d5e2031ae7120b44493c41bc WatchSource:0}: Error finding container cd22b705caa57910fe8b0cc4a2a4875140ee9107d5e2031ae7120b44493c41bc: Status 404 returned error can't find the container with id cd22b705caa57910fe8b0cc4a2a4875140ee9107d5e2031ae7120b44493c41bc Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.192817 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.237817 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.243223 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65b767dfc8-l5jkz" event={"ID":"f748658a-398b-4b6e-9f27-d2a45e6f6827","Type":"ContainerStarted","Data":"cd22b705caa57910fe8b0cc4a2a4875140ee9107d5e2031ae7120b44493c41bc"} Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.243481 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.272397 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.274994 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.302330 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.380152 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 08 15:34:42 crc kubenswrapper[4945]: I1008 15:34:42.405134 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 08 15:34:43 crc kubenswrapper[4945]: I1008 15:34:43.254068 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65b767dfc8-l5jkz" event={"ID":"f748658a-398b-4b6e-9f27-d2a45e6f6827","Type":"ContainerStarted","Data":"aef033486e083c17c7f13b7c617131bcef146adf64db9cf23beb99ea8a4700de"} Oct 08 15:34:43 crc kubenswrapper[4945]: I1008 15:34:43.261209 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 15:34:43 crc kubenswrapper[4945]: I1008 15:34:43.299911 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 08 15:34:44 crc kubenswrapper[4945]: I1008 15:34:44.263125 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:34:44 crc kubenswrapper[4945]: I1008 15:34:44.287362 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-65b767dfc8-l5jkz" podStartSLOduration=3.287336871 podStartE2EDuration="3.287336871s" podCreationTimestamp="2025-10-08 15:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:34:44.281256788 +0000 UTC m=+1293.635171699" watchObservedRunningTime="2025-10-08 15:34:44.287336871 +0000 UTC m=+1293.641251782" Oct 08 15:34:45 crc kubenswrapper[4945]: E1008 15:34:45.025561 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-glance-api:watcher_latest\\\"\"" pod="openstack/glance-db-sync-4bnbb" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.281188 4945 generic.go:334] "Generic (PLEG): container finished" podID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerID="c29850611bb832f076ebe4db8820b3e4631318cb4c5ab1a4a0f59924f416f1b1" exitCode=1 Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.281500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerDied","Data":"c29850611bb832f076ebe4db8820b3e4631318cb4c5ab1a4a0f59924f416f1b1"} Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.281532 4945 scope.go:117] "RemoveContainer" containerID="5e411c682b3d7d408df995f49aeca0a5f53e009064805c8b6559892f2c9d1713" Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.282304 4945 scope.go:117] "RemoveContainer" containerID="c29850611bb832f076ebe4db8820b3e4631318cb4c5ab1a4a0f59924f416f1b1" Oct 08 15:34:46 crc kubenswrapper[4945]: E1008 15:34:46.282566 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.489773 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.490012 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api-log" containerID="cri-o://0cce6c7dc8a496d9735089da371c1f6d51223aae0fa2bf7b46a8ecf950fd2abc" gracePeriod=30 Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.490052 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api" containerID="cri-o://b407ee8c9926a229b916162eea1d5ec04d48149647ab7b7b7974cb5fc1f2441f" gracePeriod=30 Oct 08 15:34:46 crc kubenswrapper[4945]: I1008 15:34:46.954932 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c58b7c96-vjh9q" podUID="37d6d391-e363-42db-8a0e-4e837e4af4db" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.161:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.161:8443: connect: connection refused" Oct 08 15:34:47 crc kubenswrapper[4945]: I1008 15:34:47.165913 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 08 15:34:47 crc kubenswrapper[4945]: I1008 15:34:47.274997 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9322/\": dial tcp 10.217.0.158:9322: connect: connection refused" Oct 08 15:34:47 crc kubenswrapper[4945]: I1008 15:34:47.275001 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.158:9322/\": dial tcp 10.217.0.158:9322: connect: connection refused" Oct 08 15:34:47 crc kubenswrapper[4945]: I1008 15:34:47.289968 4945 generic.go:334] "Generic (PLEG): container finished" podID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerID="0cce6c7dc8a496d9735089da371c1f6d51223aae0fa2bf7b46a8ecf950fd2abc" exitCode=143 Oct 08 15:34:47 crc kubenswrapper[4945]: I1008 15:34:47.290020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55","Type":"ContainerDied","Data":"0cce6c7dc8a496d9735089da371c1f6d51223aae0fa2bf7b46a8ecf950fd2abc"} Oct 08 15:34:48 crc kubenswrapper[4945]: I1008 15:34:48.301473 4945 generic.go:334] "Generic (PLEG): container finished" podID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerID="b407ee8c9926a229b916162eea1d5ec04d48149647ab7b7b7974cb5fc1f2441f" exitCode=0 Oct 08 15:34:48 crc kubenswrapper[4945]: I1008 15:34:48.301553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55","Type":"ContainerDied","Data":"b407ee8c9926a229b916162eea1d5ec04d48149647ab7b7b7974cb5fc1f2441f"} Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.192526 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.193986 4945 scope.go:117] "RemoveContainer" containerID="c29850611bb832f076ebe4db8820b3e4631318cb4c5ab1a4a0f59924f416f1b1" Oct 08 15:34:52 crc kubenswrapper[4945]: E1008 15:34:52.194505 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.417094 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.561322 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-combined-ca-bundle\") pod \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.561795 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlzgl\" (UniqueName: \"kubernetes.io/projected/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-kube-api-access-wlzgl\") pod \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.561862 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-custom-prometheus-ca\") pod \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.561931 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-config-data\") pod \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.561982 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-logs\") pod \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\" (UID: \"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55\") " Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.562664 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-logs" (OuterVolumeSpecName: "logs") pod "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" (UID: "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.574068 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-kube-api-access-wlzgl" (OuterVolumeSpecName: "kube-api-access-wlzgl") pod "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" (UID: "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55"). InnerVolumeSpecName "kube-api-access-wlzgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.615278 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" (UID: "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.620418 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" (UID: "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.620798 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-config-data" (OuterVolumeSpecName: "config-data") pod "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" (UID: "846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.663814 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.663872 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlzgl\" (UniqueName: \"kubernetes.io/projected/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-kube-api-access-wlzgl\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.663883 4945 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.663892 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:52 crc kubenswrapper[4945]: I1008 15:34:52.663904 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.360321 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pfbzj" event={"ID":"1be99142-5236-4365-8a45-70847b54b883","Type":"ContainerStarted","Data":"bea1ed470f3c1f66f159279f8b32dd1d7a731edce53ff621fa977d3372b1cac1"} Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.371836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ttghn" event={"ID":"0e670ed1-808d-4d6c-aa73-106476663b46","Type":"ContainerStarted","Data":"e5feed9bc5222c25f23c035f6293901350cf1dda4f5023e3a461ca8f7264faba"} Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.375220 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7fh4" event={"ID":"47794116-db67-4451-858c-bb6fea64a21c","Type":"ContainerStarted","Data":"4bb443fca884f14d549c2bf017984adacea0335b8a1163dc787f72f7f492f51a"} Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.377858 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55","Type":"ContainerDied","Data":"381123a4aedec215745f9b626e42061ef89280f7d8365266d3122165ece229e4"} Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.377921 4945 scope.go:117] "RemoveContainer" containerID="b407ee8c9926a229b916162eea1d5ec04d48149647ab7b7b7974cb5fc1f2441f" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.378023 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.383221 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-pfbzj" podStartSLOduration=2.746891255 podStartE2EDuration="56.38320474s" podCreationTimestamp="2025-10-08 15:33:57 +0000 UTC" firstStartedPulling="2025-10-08 15:33:58.44265878 +0000 UTC m=+1247.796573681" lastFinishedPulling="2025-10-08 15:34:52.078972275 +0000 UTC m=+1301.432887166" observedRunningTime="2025-10-08 15:34:53.377297342 +0000 UTC m=+1302.731212243" watchObservedRunningTime="2025-10-08 15:34:53.38320474 +0000 UTC m=+1302.737119631" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.409994 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ttghn" podStartSLOduration=2.6497793080000003 podStartE2EDuration="56.409971143s" podCreationTimestamp="2025-10-08 15:33:57 +0000 UTC" firstStartedPulling="2025-10-08 15:33:58.593405399 +0000 UTC m=+1247.947320300" lastFinishedPulling="2025-10-08 15:34:52.353597234 +0000 UTC m=+1301.707512135" observedRunningTime="2025-10-08 15:34:53.400095195 +0000 UTC m=+1302.754010096" watchObservedRunningTime="2025-10-08 15:34:53.409971143 +0000 UTC m=+1302.763886064" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.426735 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-s7fh4" podStartSLOduration=2.428806714 podStartE2EDuration="56.426698293s" podCreationTimestamp="2025-10-08 15:33:57 +0000 UTC" firstStartedPulling="2025-10-08 15:33:58.35867732 +0000 UTC m=+1247.712592221" lastFinishedPulling="2025-10-08 15:34:52.356568899 +0000 UTC m=+1301.710483800" observedRunningTime="2025-10-08 15:34:53.417881961 +0000 UTC m=+1302.771796862" watchObservedRunningTime="2025-10-08 15:34:53.426698293 +0000 UTC m=+1302.780613194" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.450817 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.475268 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.489280 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:53 crc kubenswrapper[4945]: E1008 15:34:53.489889 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.489912 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api" Oct 08 15:34:53 crc kubenswrapper[4945]: E1008 15:34:53.489947 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api-log" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.489958 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api-log" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.490237 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api-log" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.490256 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.491857 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.495102 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.498751 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.499029 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.499161 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.679809 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.679959 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-config-data\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.680058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.680439 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.680545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-logs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.680807 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-public-tls-certs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.680891 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc6lf\" (UniqueName: \"kubernetes.io/projected/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-kube-api-access-zc6lf\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.782837 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.782900 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-config-data\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.782943 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.782979 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.783029 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-logs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.783127 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-public-tls-certs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.783158 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc6lf\" (UniqueName: \"kubernetes.io/projected/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-kube-api-access-zc6lf\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.783571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-logs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.786720 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.787184 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-config-data\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.787511 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.788232 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.800391 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-public-tls-certs\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.804730 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc6lf\" (UniqueName: \"kubernetes.io/projected/f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435-kube-api-access-zc6lf\") pod \"watcher-api-0\" (UID: \"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435\") " pod="openstack/watcher-api-0" Oct 08 15:34:53 crc kubenswrapper[4945]: I1008 15:34:53.814917 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 15:34:54 crc kubenswrapper[4945]: I1008 15:34:54.043407 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" path="/var/lib/kubelet/pods/846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55/volumes" Oct 08 15:34:57 crc kubenswrapper[4945]: I1008 15:34:57.277210 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.158:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 15:34:57 crc kubenswrapper[4945]: I1008 15:34:57.277270 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="846d5a7e-cf0a-4dc2-8f3d-3efa8e571a55" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 15:34:57 crc kubenswrapper[4945]: I1008 15:34:57.595921 4945 scope.go:117] "RemoveContainer" containerID="0cce6c7dc8a496d9735089da371c1f6d51223aae0fa2bf7b46a8ecf950fd2abc" Oct 08 15:34:57 crc kubenswrapper[4945]: E1008 15:34:57.979877 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" Oct 08 15:34:58 crc kubenswrapper[4945]: W1008 15:34:58.164025 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9d2ef74_dfaf_4dc0_8af8_4b7f6b262435.slice/crio-c7d549009b5bed693ee3fde49f19898596e4c8a5c5a407980489c0ab6300e7bb WatchSource:0}: Error finding container c7d549009b5bed693ee3fde49f19898596e4c8a5c5a407980489c0ab6300e7bb: Status 404 returned error can't find the container with id c7d549009b5bed693ee3fde49f19898596e4c8a5c5a407980489c0ab6300e7bb Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.168833 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.435146 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435","Type":"ContainerStarted","Data":"2a1b4e72767523b7875e71256f7effa53471d6ade0879851281ee2bebc7516ce"} Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.435536 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435","Type":"ContainerStarted","Data":"c7d549009b5bed693ee3fde49f19898596e4c8a5c5a407980489c0ab6300e7bb"} Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.438786 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerStarted","Data":"4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179"} Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.438972 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="ceilometer-notification-agent" containerID="cri-o://071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d" gracePeriod=30 Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.439366 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.439693 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="proxy-httpd" containerID="cri-o://4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179" gracePeriod=30 Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.439768 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="sg-core" containerID="cri-o://ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32" gracePeriod=30 Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.830223 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:34:58 crc kubenswrapper[4945]: I1008 15:34:58.960299 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:34:59 crc kubenswrapper[4945]: I1008 15:34:59.457311 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435","Type":"ContainerStarted","Data":"b9d72ec801fee22d6c50ec91ea324cf5cf1ec7693c9b49f72b1f05cb5ec1c0ac"} Oct 08 15:34:59 crc kubenswrapper[4945]: I1008 15:34:59.457713 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 15:34:59 crc kubenswrapper[4945]: I1008 15:34:59.461254 4945 generic.go:334] "Generic (PLEG): container finished" podID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerID="4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179" exitCode=0 Oct 08 15:34:59 crc kubenswrapper[4945]: I1008 15:34:59.461313 4945 generic.go:334] "Generic (PLEG): container finished" podID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerID="ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32" exitCode=2 Oct 08 15:34:59 crc kubenswrapper[4945]: I1008 15:34:59.461353 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerDied","Data":"4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179"} Oct 08 15:34:59 crc kubenswrapper[4945]: I1008 15:34:59.461406 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerDied","Data":"ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32"} Oct 08 15:34:59 crc kubenswrapper[4945]: I1008 15:34:59.499793 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=6.499766133 podStartE2EDuration="6.499766133s" podCreationTimestamp="2025-10-08 15:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:34:59.490767887 +0000 UTC m=+1308.844682878" watchObservedRunningTime="2025-10-08 15:34:59.499766133 +0000 UTC m=+1308.853681074" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.390804 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.476518 4945 generic.go:334] "Generic (PLEG): container finished" podID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerID="071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d" exitCode=0 Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.477348 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.477785 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerDied","Data":"071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d"} Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.477814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86db9194-a67b-4491-b0f1-d03bafb3e353","Type":"ContainerDied","Data":"730a0ad4c6a64576b5f645e20ce116a280249b2edd000429172391f8f5357274"} Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.477830 4945 scope.go:117] "RemoveContainer" containerID="4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.498856 4945 scope.go:117] "RemoveContainer" containerID="ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.517057 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-log-httpd\") pod \"86db9194-a67b-4491-b0f1-d03bafb3e353\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.517318 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmnck\" (UniqueName: \"kubernetes.io/projected/86db9194-a67b-4491-b0f1-d03bafb3e353-kube-api-access-qmnck\") pod \"86db9194-a67b-4491-b0f1-d03bafb3e353\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.517382 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-config-data\") pod \"86db9194-a67b-4491-b0f1-d03bafb3e353\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.517455 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-sg-core-conf-yaml\") pod \"86db9194-a67b-4491-b0f1-d03bafb3e353\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.517518 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-run-httpd\") pod \"86db9194-a67b-4491-b0f1-d03bafb3e353\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.517550 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-combined-ca-bundle\") pod \"86db9194-a67b-4491-b0f1-d03bafb3e353\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.517643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-scripts\") pod \"86db9194-a67b-4491-b0f1-d03bafb3e353\" (UID: \"86db9194-a67b-4491-b0f1-d03bafb3e353\") " Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.519493 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "86db9194-a67b-4491-b0f1-d03bafb3e353" (UID: "86db9194-a67b-4491-b0f1-d03bafb3e353"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.519694 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "86db9194-a67b-4491-b0f1-d03bafb3e353" (UID: "86db9194-a67b-4491-b0f1-d03bafb3e353"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.522080 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-scripts" (OuterVolumeSpecName: "scripts") pod "86db9194-a67b-4491-b0f1-d03bafb3e353" (UID: "86db9194-a67b-4491-b0f1-d03bafb3e353"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.523201 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86db9194-a67b-4491-b0f1-d03bafb3e353-kube-api-access-qmnck" (OuterVolumeSpecName: "kube-api-access-qmnck") pod "86db9194-a67b-4491-b0f1-d03bafb3e353" (UID: "86db9194-a67b-4491-b0f1-d03bafb3e353"). InnerVolumeSpecName "kube-api-access-qmnck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.544812 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "86db9194-a67b-4491-b0f1-d03bafb3e353" (UID: "86db9194-a67b-4491-b0f1-d03bafb3e353"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.570251 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86db9194-a67b-4491-b0f1-d03bafb3e353" (UID: "86db9194-a67b-4491-b0f1-d03bafb3e353"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.594707 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-config-data" (OuterVolumeSpecName: "config-data") pod "86db9194-a67b-4491-b0f1-d03bafb3e353" (UID: "86db9194-a67b-4491-b0f1-d03bafb3e353"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.620639 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.620672 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.620681 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.620690 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.620698 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86db9194-a67b-4491-b0f1-d03bafb3e353-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.620706 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmnck\" (UniqueName: \"kubernetes.io/projected/86db9194-a67b-4491-b0f1-d03bafb3e353-kube-api-access-qmnck\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.620715 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86db9194-a67b-4491-b0f1-d03bafb3e353-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.626538 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7c58b7c96-vjh9q" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.694925 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58798f7484-jkqnb"] Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.695205 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon-log" containerID="cri-o://39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87" gracePeriod=30 Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.695326 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" containerID="cri-o://1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b" gracePeriod=30 Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.704081 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.705085 4945 scope.go:117] "RemoveContainer" containerID="071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.712329 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:42958->10.217.0.160:8443: read: connection reset by peer" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.749016 4945 scope.go:117] "RemoveContainer" containerID="4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179" Oct 08 15:35:00 crc kubenswrapper[4945]: E1008 15:35:00.750260 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179\": container with ID starting with 4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179 not found: ID does not exist" containerID="4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.750291 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179"} err="failed to get container status \"4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179\": rpc error: code = NotFound desc = could not find container \"4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179\": container with ID starting with 4640326c78ec80897fef52440c5796c390632dfc1f018a70c5b1d95bfd6f2179 not found: ID does not exist" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.750310 4945 scope.go:117] "RemoveContainer" containerID="ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32" Oct 08 15:35:00 crc kubenswrapper[4945]: E1008 15:35:00.751545 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32\": container with ID starting with ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32 not found: ID does not exist" containerID="ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.751570 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32"} err="failed to get container status \"ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32\": rpc error: code = NotFound desc = could not find container \"ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32\": container with ID starting with ba11b37d89a1507f242285aed46b375d4eeff457cf9d4e3630530b84366eac32 not found: ID does not exist" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.751584 4945 scope.go:117] "RemoveContainer" containerID="071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d" Oct 08 15:35:00 crc kubenswrapper[4945]: E1008 15:35:00.751900 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d\": container with ID starting with 071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d not found: ID does not exist" containerID="071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.751926 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d"} err="failed to get container status \"071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d\": rpc error: code = NotFound desc = could not find container \"071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d\": container with ID starting with 071409950d1acec38794987b8733a4f7bb1c802d23e9fac168875032a9be6a1d not found: ID does not exist" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.844843 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.848578 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.873338 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:00 crc kubenswrapper[4945]: E1008 15:35:00.873725 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="proxy-httpd" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.873742 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="proxy-httpd" Oct 08 15:35:00 crc kubenswrapper[4945]: E1008 15:35:00.873757 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="sg-core" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.873768 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="sg-core" Oct 08 15:35:00 crc kubenswrapper[4945]: E1008 15:35:00.873781 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="ceilometer-notification-agent" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.873788 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="ceilometer-notification-agent" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.873991 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="ceilometer-notification-agent" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.874009 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="sg-core" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.874032 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" containerName="proxy-httpd" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.897564 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.897677 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.902494 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:35:00 crc kubenswrapper[4945]: I1008 15:35:00.902799 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.029575 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-scripts\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.029612 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.029657 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4jgh\" (UniqueName: \"kubernetes.io/projected/b00bb163-6180-4acb-a485-095a289db80a-kube-api-access-g4jgh\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.029678 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-run-httpd\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.029721 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-config-data\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.029812 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-log-httpd\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.029863 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.131186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4jgh\" (UniqueName: \"kubernetes.io/projected/b00bb163-6180-4acb-a485-095a289db80a-kube-api-access-g4jgh\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.132150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-run-httpd\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.132632 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-config-data\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.132934 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-log-httpd\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.132997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.133077 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-scripts\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.133096 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.132560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-run-httpd\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.133922 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-log-httpd\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.137605 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.137906 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.138353 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-config-data\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.139426 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-scripts\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.149155 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4jgh\" (UniqueName: \"kubernetes.io/projected/b00bb163-6180-4acb-a485-095a289db80a-kube-api-access-g4jgh\") pod \"ceilometer-0\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.227630 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:01 crc kubenswrapper[4945]: E1008 15:35:01.740578 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e99e039_6910_4bca_b0ff_815eba456ebd.slice/crio-1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e99e039_6910_4bca_b0ff_815eba456ebd.slice/crio-conmon-1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b.scope\": RecentStats: unable to find data in memory cache]" Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.828761 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:01 crc kubenswrapper[4945]: I1008 15:35:01.831935 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 15:35:01 crc kubenswrapper[4945]: W1008 15:35:01.832170 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb00bb163_6180_4acb_a485_095a289db80a.slice/crio-abdb5ca063d09e7861b68b34c9a7c282768c0f44d34e5be13bf843feb873c1e7 WatchSource:0}: Error finding container abdb5ca063d09e7861b68b34c9a7c282768c0f44d34e5be13bf843feb873c1e7: Status 404 returned error can't find the container with id abdb5ca063d09e7861b68b34c9a7c282768c0f44d34e5be13bf843feb873c1e7 Oct 08 15:35:02 crc kubenswrapper[4945]: I1008 15:35:02.048063 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86db9194-a67b-4491-b0f1-d03bafb3e353" path="/var/lib/kubelet/pods/86db9194-a67b-4491-b0f1-d03bafb3e353/volumes" Oct 08 15:35:02 crc kubenswrapper[4945]: I1008 15:35:02.193696 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:02 crc kubenswrapper[4945]: I1008 15:35:02.194750 4945 scope.go:117] "RemoveContainer" containerID="c29850611bb832f076ebe4db8820b3e4631318cb4c5ab1a4a0f59924f416f1b1" Oct 08 15:35:02 crc kubenswrapper[4945]: I1008 15:35:02.197340 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:02 crc kubenswrapper[4945]: I1008 15:35:02.496150 4945 generic.go:334] "Generic (PLEG): container finished" podID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerID="1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b" exitCode=0 Oct 08 15:35:02 crc kubenswrapper[4945]: I1008 15:35:02.496215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58798f7484-jkqnb" event={"ID":"4e99e039-6910-4bca-b0ff-815eba456ebd","Type":"ContainerDied","Data":"1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b"} Oct 08 15:35:02 crc kubenswrapper[4945]: I1008 15:35:02.497441 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerStarted","Data":"abdb5ca063d09e7861b68b34c9a7c282768c0f44d34e5be13bf843feb873c1e7"} Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.508756 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerStarted","Data":"a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6"} Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.510896 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4bnbb" event={"ID":"0646f7b0-6abf-436c-976f-59ab119525ad","Type":"ContainerStarted","Data":"d51a29b99688467a7a64d40bba0f998eb8efc47d4f0e03d26b39c3dc69ab162f"} Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.513243 4945 generic.go:334] "Generic (PLEG): container finished" podID="47794116-db67-4451-858c-bb6fea64a21c" containerID="4bb443fca884f14d549c2bf017984adacea0335b8a1163dc787f72f7f492f51a" exitCode=0 Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.513295 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7fh4" event={"ID":"47794116-db67-4451-858c-bb6fea64a21c","Type":"ContainerDied","Data":"4bb443fca884f14d549c2bf017984adacea0335b8a1163dc787f72f7f492f51a"} Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.515395 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerStarted","Data":"8395ee59d8bef1a2e368ede09a30682e3b96e97a08837fe250471dad4cb35de5"} Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.515456 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerStarted","Data":"32502233dfa60024f099b1e52bd8769c232d782ad1560a3c56753539d5a5f6f7"} Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.564578 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4bnbb" podStartSLOduration=4.302814653 podStartE2EDuration="1m24.56455083s" podCreationTimestamp="2025-10-08 15:33:39 +0000 UTC" firstStartedPulling="2025-10-08 15:33:39.915479185 +0000 UTC m=+1229.269394086" lastFinishedPulling="2025-10-08 15:35:00.177215362 +0000 UTC m=+1309.531130263" observedRunningTime="2025-10-08 15:35:03.554774584 +0000 UTC m=+1312.908689485" watchObservedRunningTime="2025-10-08 15:35:03.56455083 +0000 UTC m=+1312.918465771" Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.825204 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.825554 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 15:35:03 crc kubenswrapper[4945]: I1008 15:35:03.832898 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 08 15:35:04 crc kubenswrapper[4945]: I1008 15:35:04.069782 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:35:04 crc kubenswrapper[4945]: I1008 15:35:04.523836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerStarted","Data":"279951b5df62c3b8c95708382849b0f083ceb676b2e74662684f60764ce3c250"} Oct 08 15:35:04 crc kubenswrapper[4945]: I1008 15:35:04.525326 4945 generic.go:334] "Generic (PLEG): container finished" podID="0e670ed1-808d-4d6c-aa73-106476663b46" containerID="e5feed9bc5222c25f23c035f6293901350cf1dda4f5023e3a461ca8f7264faba" exitCode=0 Oct 08 15:35:04 crc kubenswrapper[4945]: I1008 15:35:04.525503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ttghn" event={"ID":"0e670ed1-808d-4d6c-aa73-106476663b46","Type":"ContainerDied","Data":"e5feed9bc5222c25f23c035f6293901350cf1dda4f5023e3a461ca8f7264faba"} Oct 08 15:35:04 crc kubenswrapper[4945]: I1008 15:35:04.532565 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 15:35:04 crc kubenswrapper[4945]: I1008 15:35:04.930786 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7fh4" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.110056 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbnnp\" (UniqueName: \"kubernetes.io/projected/47794116-db67-4451-858c-bb6fea64a21c-kube-api-access-lbnnp\") pod \"47794116-db67-4451-858c-bb6fea64a21c\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.110165 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47794116-db67-4451-858c-bb6fea64a21c-logs\") pod \"47794116-db67-4451-858c-bb6fea64a21c\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.110252 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-config-data\") pod \"47794116-db67-4451-858c-bb6fea64a21c\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.110301 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-combined-ca-bundle\") pod \"47794116-db67-4451-858c-bb6fea64a21c\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.110400 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-scripts\") pod \"47794116-db67-4451-858c-bb6fea64a21c\" (UID: \"47794116-db67-4451-858c-bb6fea64a21c\") " Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.110689 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47794116-db67-4451-858c-bb6fea64a21c-logs" (OuterVolumeSpecName: "logs") pod "47794116-db67-4451-858c-bb6fea64a21c" (UID: "47794116-db67-4451-858c-bb6fea64a21c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.110831 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47794116-db67-4451-858c-bb6fea64a21c-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.116935 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-scripts" (OuterVolumeSpecName: "scripts") pod "47794116-db67-4451-858c-bb6fea64a21c" (UID: "47794116-db67-4451-858c-bb6fea64a21c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.117300 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47794116-db67-4451-858c-bb6fea64a21c-kube-api-access-lbnnp" (OuterVolumeSpecName: "kube-api-access-lbnnp") pod "47794116-db67-4451-858c-bb6fea64a21c" (UID: "47794116-db67-4451-858c-bb6fea64a21c"). InnerVolumeSpecName "kube-api-access-lbnnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.152539 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47794116-db67-4451-858c-bb6fea64a21c" (UID: "47794116-db67-4451-858c-bb6fea64a21c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.156585 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-config-data" (OuterVolumeSpecName: "config-data") pod "47794116-db67-4451-858c-bb6fea64a21c" (UID: "47794116-db67-4451-858c-bb6fea64a21c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.213053 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbnnp\" (UniqueName: \"kubernetes.io/projected/47794116-db67-4451-858c-bb6fea64a21c-kube-api-access-lbnnp\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.213090 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.213149 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.213169 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47794116-db67-4451-858c-bb6fea64a21c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.536172 4945 generic.go:334] "Generic (PLEG): container finished" podID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerID="a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6" exitCode=1 Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.536263 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerDied","Data":"a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6"} Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.536591 4945 scope.go:117] "RemoveContainer" containerID="c29850611bb832f076ebe4db8820b3e4631318cb4c5ab1a4a0f59924f416f1b1" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.537319 4945 scope.go:117] "RemoveContainer" containerID="a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6" Oct 08 15:35:05 crc kubenswrapper[4945]: E1008 15:35:05.537759 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.539620 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7fh4" event={"ID":"47794116-db67-4451-858c-bb6fea64a21c","Type":"ContainerDied","Data":"b854e7f15557e6a4631c07fa185cbdcf9c7ee4742b883fd5d7e0a5c89f5e0ada"} Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.539727 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b854e7f15557e6a4631c07fa185cbdcf9c7ee4742b883fd5d7e0a5c89f5e0ada" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.539816 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7fh4" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.743970 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-66d5fb75f8-44p9s"] Oct 08 15:35:05 crc kubenswrapper[4945]: E1008 15:35:05.744395 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47794116-db67-4451-858c-bb6fea64a21c" containerName="placement-db-sync" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.744408 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="47794116-db67-4451-858c-bb6fea64a21c" containerName="placement-db-sync" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.744587 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="47794116-db67-4451-858c-bb6fea64a21c" containerName="placement-db-sync" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.749424 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.757100 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.757286 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.761757 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tfz6r" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.762077 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.765644 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.772260 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66d5fb75f8-44p9s"] Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.930686 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66fdg\" (UniqueName: \"kubernetes.io/projected/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-kube-api-access-66fdg\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.930998 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-scripts\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.931025 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-config-data\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.931081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-logs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.931440 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-public-tls-certs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.931552 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-combined-ca-bundle\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.931574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-internal-tls-certs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:05 crc kubenswrapper[4945]: I1008 15:35:05.933078 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ttghn" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.032610 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-combined-ca-bundle\") pod \"0e670ed1-808d-4d6c-aa73-106476663b46\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.032730 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ntfm\" (UniqueName: \"kubernetes.io/projected/0e670ed1-808d-4d6c-aa73-106476663b46-kube-api-access-5ntfm\") pod \"0e670ed1-808d-4d6c-aa73-106476663b46\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.032798 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-db-sync-config-data\") pod \"0e670ed1-808d-4d6c-aa73-106476663b46\" (UID: \"0e670ed1-808d-4d6c-aa73-106476663b46\") " Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033034 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-combined-ca-bundle\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033057 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-internal-tls-certs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66fdg\" (UniqueName: \"kubernetes.io/projected/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-kube-api-access-66fdg\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033100 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-scripts\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033145 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-config-data\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033198 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-logs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-public-tls-certs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.033921 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-logs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.038258 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0e670ed1-808d-4d6c-aa73-106476663b46" (UID: "0e670ed1-808d-4d6c-aa73-106476663b46"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.038710 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-scripts\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.038928 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-combined-ca-bundle\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.039491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-internal-tls-certs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.040042 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-config-data\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.040485 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-public-tls-certs\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.043357 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e670ed1-808d-4d6c-aa73-106476663b46-kube-api-access-5ntfm" (OuterVolumeSpecName: "kube-api-access-5ntfm") pod "0e670ed1-808d-4d6c-aa73-106476663b46" (UID: "0e670ed1-808d-4d6c-aa73-106476663b46"). InnerVolumeSpecName "kube-api-access-5ntfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.060158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66fdg\" (UniqueName: \"kubernetes.io/projected/bac08d54-03b7-4701-b9c7-69cc37c2eb3d-kube-api-access-66fdg\") pod \"placement-66d5fb75f8-44p9s\" (UID: \"bac08d54-03b7-4701-b9c7-69cc37c2eb3d\") " pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.071244 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e670ed1-808d-4d6c-aa73-106476663b46" (UID: "0e670ed1-808d-4d6c-aa73-106476663b46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.114070 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.135017 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.135321 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ntfm\" (UniqueName: \"kubernetes.io/projected/0e670ed1-808d-4d6c-aa73-106476663b46-kube-api-access-5ntfm\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.135420 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0e670ed1-808d-4d6c-aa73-106476663b46-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.551155 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ttghn" event={"ID":"0e670ed1-808d-4d6c-aa73-106476663b46","Type":"ContainerDied","Data":"ca847cdf5fa40e87a3bdc5754354c37bace6da9a51dfba4244a7dd53edd4e0bb"} Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.551534 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca847cdf5fa40e87a3bdc5754354c37bace6da9a51dfba4244a7dd53edd4e0bb" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.551175 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ttghn" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.555449 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerStarted","Data":"2f5f9d1ad32e9a7ee9cfa649f015b3c74f56800032144d01bcedd0537b6c0114"} Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.555914 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.573320 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66d5fb75f8-44p9s"] Oct 08 15:35:06 crc kubenswrapper[4945]: W1008 15:35:06.583778 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbac08d54_03b7_4701_b9c7_69cc37c2eb3d.slice/crio-ea21ef6252032de3e7d23d1cc4e0fe03297dc2db0c512324f74cc2557c8f0e40 WatchSource:0}: Error finding container ea21ef6252032de3e7d23d1cc4e0fe03297dc2db0c512324f74cc2557c8f0e40: Status 404 returned error can't find the container with id ea21ef6252032de3e7d23d1cc4e0fe03297dc2db0c512324f74cc2557c8f0e40 Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.599289 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.775677451 podStartE2EDuration="6.599266669s" podCreationTimestamp="2025-10-08 15:35:00 +0000 UTC" firstStartedPulling="2025-10-08 15:35:01.835216905 +0000 UTC m=+1311.189131806" lastFinishedPulling="2025-10-08 15:35:05.658806123 +0000 UTC m=+1315.012721024" observedRunningTime="2025-10-08 15:35:06.581626795 +0000 UTC m=+1315.935541696" watchObservedRunningTime="2025-10-08 15:35:06.599266669 +0000 UTC m=+1315.953181570" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.815693 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-fc897d669-lcbrz"] Oct 08 15:35:06 crc kubenswrapper[4945]: E1008 15:35:06.816384 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e670ed1-808d-4d6c-aa73-106476663b46" containerName="barbican-db-sync" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.816404 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e670ed1-808d-4d6c-aa73-106476663b46" containerName="barbican-db-sync" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.816627 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e670ed1-808d-4d6c-aa73-106476663b46" containerName="barbican-db-sync" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.817586 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.823096 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.823567 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.823789 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kqbsg" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.853159 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-b8d567b6d-zc6xh"] Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.861650 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.876494 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.878582 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-combined-ca-bundle\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.878641 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfqdz\" (UniqueName: \"kubernetes.io/projected/dfcd7c98-f494-4669-8e76-995ccb3e768b-kube-api-access-nfqdz\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.878687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-config-data-custom\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.878741 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfcd7c98-f494-4669-8e76-995ccb3e768b-logs\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.878776 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-config-data\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.885941 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fc897d669-lcbrz"] Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.970278 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b8d567b6d-zc6xh"] Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983736 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-config-data-custom\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983789 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-combined-ca-bundle\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983826 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh58f\" (UniqueName: \"kubernetes.io/projected/66b89671-2c42-4e97-8a74-0e03bac0613a-kube-api-access-xh58f\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983855 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfqdz\" (UniqueName: \"kubernetes.io/projected/dfcd7c98-f494-4669-8e76-995ccb3e768b-kube-api-access-nfqdz\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983888 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-combined-ca-bundle\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983914 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-config-data-custom\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983965 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfcd7c98-f494-4669-8e76-995ccb3e768b-logs\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.983986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b89671-2c42-4e97-8a74-0e03bac0613a-logs\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.984008 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-config-data\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.984549 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-config-data\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.987594 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfcd7c98-f494-4669-8e76-995ccb3e768b-logs\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:06 crc kubenswrapper[4945]: I1008 15:35:06.994257 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-config-data-custom\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.005887 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-config-data\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.009584 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfcd7c98-f494-4669-8e76-995ccb3e768b-combined-ca-bundle\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.012521 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfqdz\" (UniqueName: \"kubernetes.io/projected/dfcd7c98-f494-4669-8e76-995ccb3e768b-kube-api-access-nfqdz\") pod \"barbican-worker-fc897d669-lcbrz\" (UID: \"dfcd7c98-f494-4669-8e76-995ccb3e768b\") " pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.023014 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66898c8f85-lp5qv"] Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.030752 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.064902 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66898c8f85-lp5qv"] Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.074048 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c469d5fc4-4hmvj"] Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.076027 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.078305 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089400 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-svc\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-config-data-custom\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089490 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh58f\" (UniqueName: \"kubernetes.io/projected/66b89671-2c42-4e97-8a74-0e03bac0613a-kube-api-access-xh58f\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089517 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-swift-storage-0\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089547 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-combined-ca-bundle\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-config\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089656 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-665vc\" (UniqueName: \"kubernetes.io/projected/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-kube-api-access-665vc\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089677 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b89671-2c42-4e97-8a74-0e03bac0613a-logs\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-nb\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089718 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-sb\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.089744 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-config-data\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.099671 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-config-data-custom\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.099946 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b89671-2c42-4e97-8a74-0e03bac0613a-logs\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.110452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-combined-ca-bundle\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.126153 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b89671-2c42-4e97-8a74-0e03bac0613a-config-data\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.133163 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c469d5fc4-4hmvj"] Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.150814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh58f\" (UniqueName: \"kubernetes.io/projected/66b89671-2c42-4e97-8a74-0e03bac0613a-kube-api-access-xh58f\") pod \"barbican-keystone-listener-b8d567b6d-zc6xh\" (UID: \"66b89671-2c42-4e97-8a74-0e03bac0613a\") " pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.173248 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191649 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-config\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ljgl\" (UniqueName: \"kubernetes.io/projected/604863fd-6838-43ef-96a9-393af1ec3c2e-kube-api-access-9ljgl\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-665vc\" (UniqueName: \"kubernetes.io/projected/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-kube-api-access-665vc\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191781 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-combined-ca-bundle\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191805 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191825 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604863fd-6838-43ef-96a9-393af1ec3c2e-logs\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191847 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-nb\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191864 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-sb\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191896 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-svc\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191948 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-swift-storage-0\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.191965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data-custom\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.193077 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-config\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.193924 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-sb\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.194552 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-nb\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.194927 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-swift-storage-0\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.201777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-svc\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.229900 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-665vc\" (UniqueName: \"kubernetes.io/projected/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-kube-api-access-665vc\") pod \"dnsmasq-dns-66898c8f85-lp5qv\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.238564 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fc897d669-lcbrz" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.296224 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ljgl\" (UniqueName: \"kubernetes.io/projected/604863fd-6838-43ef-96a9-393af1ec3c2e-kube-api-access-9ljgl\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.296297 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-combined-ca-bundle\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.296337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.296368 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604863fd-6838-43ef-96a9-393af1ec3c2e-logs\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.296524 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data-custom\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.301411 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604863fd-6838-43ef-96a9-393af1ec3c2e-logs\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.304233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.308194 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-combined-ca-bundle\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.322280 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data-custom\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.328838 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ljgl\" (UniqueName: \"kubernetes.io/projected/604863fd-6838-43ef-96a9-393af1ec3c2e-kube-api-access-9ljgl\") pod \"barbican-api-c469d5fc4-4hmvj\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.366415 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.510703 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.533927 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.592983 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66d5fb75f8-44p9s" event={"ID":"bac08d54-03b7-4701-b9c7-69cc37c2eb3d","Type":"ContainerStarted","Data":"f0aaf7e26e63394f7d2aceeadd000d11c21eb230e78946aa5149a8488f340231"} Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.593020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66d5fb75f8-44p9s" event={"ID":"bac08d54-03b7-4701-b9c7-69cc37c2eb3d","Type":"ContainerStarted","Data":"fdcdd18f621d2af44b0429ce1dce755eada3041706d3e16a9f231fdcfbee3b60"} Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.593030 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66d5fb75f8-44p9s" event={"ID":"bac08d54-03b7-4701-b9c7-69cc37c2eb3d","Type":"ContainerStarted","Data":"ea21ef6252032de3e7d23d1cc4e0fe03297dc2db0c512324f74cc2557c8f0e40"} Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.593056 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.593072 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.635730 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-66d5fb75f8-44p9s" podStartSLOduration=2.635693805 podStartE2EDuration="2.635693805s" podCreationTimestamp="2025-10-08 15:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:07.619565879 +0000 UTC m=+1316.973480790" watchObservedRunningTime="2025-10-08 15:35:07.635693805 +0000 UTC m=+1316.989608706" Oct 08 15:35:07 crc kubenswrapper[4945]: I1008 15:35:07.872824 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fc897d669-lcbrz"] Oct 08 15:35:07 crc kubenswrapper[4945]: W1008 15:35:07.896475 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfcd7c98_f494_4669_8e76_995ccb3e768b.slice/crio-24b4e992a5d16fbf8135c2317e67f55f0f3cb88d2ecac95f1d92b230757a29c0 WatchSource:0}: Error finding container 24b4e992a5d16fbf8135c2317e67f55f0f3cb88d2ecac95f1d92b230757a29c0: Status 404 returned error can't find the container with id 24b4e992a5d16fbf8135c2317e67f55f0f3cb88d2ecac95f1d92b230757a29c0 Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.045804 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b8d567b6d-zc6xh"] Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.140892 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66898c8f85-lp5qv"] Oct 08 15:35:08 crc kubenswrapper[4945]: W1008 15:35:08.284241 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604863fd_6838_43ef_96a9_393af1ec3c2e.slice/crio-78847e284c5fb3ad8501f7810c75e5c23810856820ca8c8eaa83ffa8d067ceae WatchSource:0}: Error finding container 78847e284c5fb3ad8501f7810c75e5c23810856820ca8c8eaa83ffa8d067ceae: Status 404 returned error can't find the container with id 78847e284c5fb3ad8501f7810c75e5c23810856820ca8c8eaa83ffa8d067ceae Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.289022 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c469d5fc4-4hmvj"] Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.605181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" event={"ID":"66b89671-2c42-4e97-8a74-0e03bac0613a","Type":"ContainerStarted","Data":"0cf67679ab0e70abedf7b400de17458d6b230adfd478761dbc8a0c8ef586c1c5"} Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.608877 4945 generic.go:334] "Generic (PLEG): container finished" podID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerID="b497f9b12634110643fbf43f63ed784c2a55eb797ab9d0cdd22e91df4c713458" exitCode=0 Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.608937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" event={"ID":"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d","Type":"ContainerDied","Data":"b497f9b12634110643fbf43f63ed784c2a55eb797ab9d0cdd22e91df4c713458"} Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.608962 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" event={"ID":"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d","Type":"ContainerStarted","Data":"7adaccee56d76c8a0ed018de9109b89b3c3f7b70549f1e5b3811e72f18fce2f6"} Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.612642 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c469d5fc4-4hmvj" event={"ID":"604863fd-6838-43ef-96a9-393af1ec3c2e","Type":"ContainerStarted","Data":"4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5"} Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.612675 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c469d5fc4-4hmvj" event={"ID":"604863fd-6838-43ef-96a9-393af1ec3c2e","Type":"ContainerStarted","Data":"78847e284c5fb3ad8501f7810c75e5c23810856820ca8c8eaa83ffa8d067ceae"} Oct 08 15:35:08 crc kubenswrapper[4945]: I1008 15:35:08.616790 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fc897d669-lcbrz" event={"ID":"dfcd7c98-f494-4669-8e76-995ccb3e768b","Type":"ContainerStarted","Data":"24b4e992a5d16fbf8135c2317e67f55f0f3cb88d2ecac95f1d92b230757a29c0"} Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.636498 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c469d5fc4-4hmvj" event={"ID":"604863fd-6838-43ef-96a9-393af1ec3c2e","Type":"ContainerStarted","Data":"886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc"} Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.659307 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c469d5fc4-4hmvj" podStartSLOduration=3.659286452 podStartE2EDuration="3.659286452s" podCreationTimestamp="2025-10-08 15:35:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:09.656921972 +0000 UTC m=+1319.010836873" watchObservedRunningTime="2025-10-08 15:35:09.659286452 +0000 UTC m=+1319.013201363" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.741983 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5588d6b8d6-7nmw2"] Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.747057 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.751568 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.751760 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.759949 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5588d6b8d6-7nmw2"] Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.855650 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-config-data\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.855747 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqf2w\" (UniqueName: \"kubernetes.io/projected/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-kube-api-access-dqf2w\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.855827 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-public-tls-certs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.855892 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-internal-tls-certs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.855960 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-logs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.856004 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-combined-ca-bundle\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.856124 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-config-data-custom\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.957614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-config-data\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.957705 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqf2w\" (UniqueName: \"kubernetes.io/projected/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-kube-api-access-dqf2w\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.957748 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-public-tls-certs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.957788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-internal-tls-certs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.957831 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-logs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.957866 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-combined-ca-bundle\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.957914 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-config-data-custom\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.958241 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-logs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.970270 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-internal-tls-certs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.971657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-public-tls-certs\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.972359 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-config-data\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.977787 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-combined-ca-bundle\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.983609 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqf2w\" (UniqueName: \"kubernetes.io/projected/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-kube-api-access-dqf2w\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:09 crc kubenswrapper[4945]: I1008 15:35:09.984061 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6-config-data-custom\") pod \"barbican-api-5588d6b8d6-7nmw2\" (UID: \"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6\") " pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:10 crc kubenswrapper[4945]: I1008 15:35:10.076667 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:10 crc kubenswrapper[4945]: I1008 15:35:10.655002 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:10 crc kubenswrapper[4945]: I1008 15:35:10.655412 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.319385 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5588d6b8d6-7nmw2"] Oct 08 15:35:11 crc kubenswrapper[4945]: W1008 15:35:11.322599 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebe5d0d3_f41d_4225_8b84_b8c16ee30ca6.slice/crio-60b134ad45d63d0f5a2085bc62ca04b41e8b11db8aca69f84d3c01ff8ee6a23c WatchSource:0}: Error finding container 60b134ad45d63d0f5a2085bc62ca04b41e8b11db8aca69f84d3c01ff8ee6a23c: Status 404 returned error can't find the container with id 60b134ad45d63d0f5a2085bc62ca04b41e8b11db8aca69f84d3c01ff8ee6a23c Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.663913 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fc897d669-lcbrz" event={"ID":"dfcd7c98-f494-4669-8e76-995ccb3e768b","Type":"ContainerStarted","Data":"ee94c7d41910c0d63e53cd2dc14577b06127809fdd94b137515dde0079b93b04"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.663956 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fc897d669-lcbrz" event={"ID":"dfcd7c98-f494-4669-8e76-995ccb3e768b","Type":"ContainerStarted","Data":"5867258da67cc0d3b5bea9a54a5366d1aa3fe4a999c8e09182a4c2b84549f3ea"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.675307 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" event={"ID":"66b89671-2c42-4e97-8a74-0e03bac0613a","Type":"ContainerStarted","Data":"c04585592127ea91cdf91ceb0b185708d16c179b08c07d21e839af7836db550e"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.675358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" event={"ID":"66b89671-2c42-4e97-8a74-0e03bac0613a","Type":"ContainerStarted","Data":"7c4bf04da4aab4c242d6b4531625f5d64d77e8ac349f520f0d542dd6cae29b50"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.683709 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" event={"ID":"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d","Type":"ContainerStarted","Data":"d9cc77780980160bead7c1d8586ff3e120885d1660fbbbb5478dce92676aeb65"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.684543 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.685553 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-fc897d669-lcbrz" podStartSLOduration=2.725945245 podStartE2EDuration="5.685530926s" podCreationTimestamp="2025-10-08 15:35:06 +0000 UTC" firstStartedPulling="2025-10-08 15:35:07.898789794 +0000 UTC m=+1317.252704685" lastFinishedPulling="2025-10-08 15:35:10.858375475 +0000 UTC m=+1320.212290366" observedRunningTime="2025-10-08 15:35:11.682003007 +0000 UTC m=+1321.035917918" watchObservedRunningTime="2025-10-08 15:35:11.685530926 +0000 UTC m=+1321.039445837" Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.689281 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5588d6b8d6-7nmw2" event={"ID":"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6","Type":"ContainerStarted","Data":"af88c6eccb197648bb44a58d9b24c5f3dda6938e849c6243c2361ab9f63f0c9d"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.689324 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5588d6b8d6-7nmw2" event={"ID":"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6","Type":"ContainerStarted","Data":"193cde49086b247ea84f458c0454a44fd3ff470a4d9a7ac2842fa8e6e8bd56b9"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.689365 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5588d6b8d6-7nmw2" event={"ID":"ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6","Type":"ContainerStarted","Data":"60b134ad45d63d0f5a2085bc62ca04b41e8b11db8aca69f84d3c01ff8ee6a23c"} Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.690001 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.690116 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.722427 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-b8d567b6d-zc6xh" podStartSLOduration=2.8841292579999998 podStartE2EDuration="5.722400962s" podCreationTimestamp="2025-10-08 15:35:06 +0000 UTC" firstStartedPulling="2025-10-08 15:35:08.047358177 +0000 UTC m=+1317.401273078" lastFinishedPulling="2025-10-08 15:35:10.885629881 +0000 UTC m=+1320.239544782" observedRunningTime="2025-10-08 15:35:11.70958692 +0000 UTC m=+1321.063501821" watchObservedRunningTime="2025-10-08 15:35:11.722400962 +0000 UTC m=+1321.076315883" Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.748605 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5588d6b8d6-7nmw2" podStartSLOduration=2.7485841300000002 podStartE2EDuration="2.74858413s" podCreationTimestamp="2025-10-08 15:35:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:11.746427946 +0000 UTC m=+1321.100342847" watchObservedRunningTime="2025-10-08 15:35:11.74858413 +0000 UTC m=+1321.102499031" Oct 08 15:35:11 crc kubenswrapper[4945]: I1008 15:35:11.778297 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" podStartSLOduration=5.778274715 podStartE2EDuration="5.778274715s" podCreationTimestamp="2025-10-08 15:35:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:11.765545176 +0000 UTC m=+1321.119460077" watchObservedRunningTime="2025-10-08 15:35:11.778274715 +0000 UTC m=+1321.132189616" Oct 08 15:35:12 crc kubenswrapper[4945]: I1008 15:35:12.192748 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:12 crc kubenswrapper[4945]: I1008 15:35:12.193054 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:12 crc kubenswrapper[4945]: I1008 15:35:12.193773 4945 scope.go:117] "RemoveContainer" containerID="a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6" Oct 08 15:35:12 crc kubenswrapper[4945]: E1008 15:35:12.193993 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:35:12 crc kubenswrapper[4945]: I1008 15:35:12.700167 4945 generic.go:334] "Generic (PLEG): container finished" podID="1be99142-5236-4365-8a45-70847b54b883" containerID="bea1ed470f3c1f66f159279f8b32dd1d7a731edce53ff621fa977d3372b1cac1" exitCode=0 Oct 08 15:35:12 crc kubenswrapper[4945]: I1008 15:35:12.700256 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pfbzj" event={"ID":"1be99142-5236-4365-8a45-70847b54b883","Type":"ContainerDied","Data":"bea1ed470f3c1f66f159279f8b32dd1d7a731edce53ff621fa977d3372b1cac1"} Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.132208 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.245639 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-scripts\") pod \"1be99142-5236-4365-8a45-70847b54b883\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.245767 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-config-data\") pod \"1be99142-5236-4365-8a45-70847b54b883\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.245838 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-combined-ca-bundle\") pod \"1be99142-5236-4365-8a45-70847b54b883\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.245942 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pclt8\" (UniqueName: \"kubernetes.io/projected/1be99142-5236-4365-8a45-70847b54b883-kube-api-access-pclt8\") pod \"1be99142-5236-4365-8a45-70847b54b883\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.245994 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1be99142-5236-4365-8a45-70847b54b883-etc-machine-id\") pod \"1be99142-5236-4365-8a45-70847b54b883\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.246033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-db-sync-config-data\") pod \"1be99142-5236-4365-8a45-70847b54b883\" (UID: \"1be99142-5236-4365-8a45-70847b54b883\") " Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.252701 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1be99142-5236-4365-8a45-70847b54b883-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1be99142-5236-4365-8a45-70847b54b883" (UID: "1be99142-5236-4365-8a45-70847b54b883"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.260891 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1be99142-5236-4365-8a45-70847b54b883-kube-api-access-pclt8" (OuterVolumeSpecName: "kube-api-access-pclt8") pod "1be99142-5236-4365-8a45-70847b54b883" (UID: "1be99142-5236-4365-8a45-70847b54b883"). InnerVolumeSpecName "kube-api-access-pclt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.260994 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1be99142-5236-4365-8a45-70847b54b883" (UID: "1be99142-5236-4365-8a45-70847b54b883"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.277821 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-scripts" (OuterVolumeSpecName: "scripts") pod "1be99142-5236-4365-8a45-70847b54b883" (UID: "1be99142-5236-4365-8a45-70847b54b883"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.305772 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1be99142-5236-4365-8a45-70847b54b883" (UID: "1be99142-5236-4365-8a45-70847b54b883"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.339065 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-config-data" (OuterVolumeSpecName: "config-data") pod "1be99142-5236-4365-8a45-70847b54b883" (UID: "1be99142-5236-4365-8a45-70847b54b883"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.348193 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pclt8\" (UniqueName: \"kubernetes.io/projected/1be99142-5236-4365-8a45-70847b54b883-kube-api-access-pclt8\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.348224 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1be99142-5236-4365-8a45-70847b54b883-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.348233 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.348242 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.348251 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.348259 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be99142-5236-4365-8a45-70847b54b883-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.577765 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-65b767dfc8-l5jkz" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.721303 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pfbzj" event={"ID":"1be99142-5236-4365-8a45-70847b54b883","Type":"ContainerDied","Data":"22fae66bba0892d3555bea57b50948363a074bc1b0dcb692063a6c5f2b8446bd"} Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.721337 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22fae66bba0892d3555bea57b50948363a074bc1b0dcb692063a6c5f2b8446bd" Oct 08 15:35:14 crc kubenswrapper[4945]: I1008 15:35:14.721402 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pfbzj" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.036144 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:15 crc kubenswrapper[4945]: E1008 15:35:15.036730 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1be99142-5236-4365-8a45-70847b54b883" containerName="cinder-db-sync" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.036746 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1be99142-5236-4365-8a45-70847b54b883" containerName="cinder-db-sync" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.036942 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1be99142-5236-4365-8a45-70847b54b883" containerName="cinder-db-sync" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.037936 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.048151 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.058416 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.058797 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.058953 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.061865 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fkwqt" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.164599 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.165166 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62c8d55c-c623-4d14-8117-91da2634ad6b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.165327 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljdz\" (UniqueName: \"kubernetes.io/projected/62c8d55c-c623-4d14-8117-91da2634ad6b-kube-api-access-8ljdz\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.165478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.165635 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-scripts\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.165745 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.168815 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66898c8f85-lp5qv"] Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.169063 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" podUID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerName="dnsmasq-dns" containerID="cri-o://d9cc77780980160bead7c1d8586ff3e120885d1660fbbbb5478dce92676aeb65" gracePeriod=10 Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.202222 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d8787f869-sll46"] Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.204625 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.213608 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8787f869-sll46"] Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.268129 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.268273 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.268345 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62c8d55c-c623-4d14-8117-91da2634ad6b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.268383 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljdz\" (UniqueName: \"kubernetes.io/projected/62c8d55c-c623-4d14-8117-91da2634ad6b-kube-api-access-8ljdz\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.268432 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.268471 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-scripts\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.269028 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62c8d55c-c623-4d14-8117-91da2634ad6b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.272558 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.272871 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-scripts\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.276768 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.292658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljdz\" (UniqueName: \"kubernetes.io/projected/62c8d55c-c623-4d14-8117-91da2634ad6b-kube-api-access-8ljdz\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.298969 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data\") pod \"cinder-scheduler-0\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.370096 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.370202 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.370272 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck7xk\" (UniqueName: \"kubernetes.io/projected/ce3b16c0-752d-4c25-9565-383b361391de-kube-api-access-ck7xk\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.370368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-config\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.370401 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-svc\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.370426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.372563 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.443624 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.449814 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.452566 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.455591 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.475751 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-config\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.475806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-svc\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.475839 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.475952 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.476004 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.476061 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck7xk\" (UniqueName: \"kubernetes.io/projected/ce3b16c0-752d-4c25-9565-383b361391de-kube-api-access-ck7xk\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.477446 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-config\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.478141 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-svc\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.478846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.479716 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.486529 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.537042 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck7xk\" (UniqueName: \"kubernetes.io/projected/ce3b16c0-752d-4c25-9565-383b361391de-kube-api-access-ck7xk\") pod \"dnsmasq-dns-5d8787f869-sll46\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.538120 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.587014 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.587176 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d41d3b-2a80-42be-b829-f627d3225331-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.587294 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.587403 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6bcb\" (UniqueName: \"kubernetes.io/projected/b6d41d3b-2a80-42be-b829-f627d3225331-kube-api-access-m6bcb\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.587509 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d41d3b-2a80-42be-b829-f627d3225331-logs\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.587657 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.587758 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-scripts\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691197 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691597 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6bcb\" (UniqueName: \"kubernetes.io/projected/b6d41d3b-2a80-42be-b829-f627d3225331-kube-api-access-m6bcb\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691674 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d41d3b-2a80-42be-b829-f627d3225331-logs\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691715 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691773 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-scripts\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691798 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691823 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d41d3b-2a80-42be-b829-f627d3225331-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.691930 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d41d3b-2a80-42be-b829-f627d3225331-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.699310 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.700305 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.700494 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d41d3b-2a80-42be-b829-f627d3225331-logs\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.702782 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-scripts\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.706229 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.729306 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6bcb\" (UniqueName: \"kubernetes.io/projected/b6d41d3b-2a80-42be-b829-f627d3225331-kube-api-access-m6bcb\") pod \"cinder-api-0\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.745980 4945 generic.go:334] "Generic (PLEG): container finished" podID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerID="d9cc77780980160bead7c1d8586ff3e120885d1660fbbbb5478dce92676aeb65" exitCode=0 Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.746048 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" event={"ID":"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d","Type":"ContainerDied","Data":"d9cc77780980160bead7c1d8586ff3e120885d1660fbbbb5478dce92676aeb65"} Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.786626 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 15:35:15 crc kubenswrapper[4945]: I1008 15:35:15.949514 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:16 crc kubenswrapper[4945]: I1008 15:35:16.324272 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8787f869-sll46"] Oct 08 15:35:16 crc kubenswrapper[4945]: I1008 15:35:16.783728 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:16 crc kubenswrapper[4945]: I1008 15:35:16.784099 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8787f869-sll46" event={"ID":"ce3b16c0-752d-4c25-9565-383b361391de","Type":"ContainerStarted","Data":"eba61c560b4954b094d617dc475adb2d764975645768c36c7d671f7c981f7119"} Oct 08 15:35:16 crc kubenswrapper[4945]: I1008 15:35:16.833584 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62c8d55c-c623-4d14-8117-91da2634ad6b","Type":"ContainerStarted","Data":"26158ba8609095114810fd0d2879e0e56a5115c958352b0ccf3d0aa5c49d29c4"} Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.168602 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.248955 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.492907 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.654121 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-swift-storage-0\") pod \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.654368 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-sb\") pod \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.654439 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-nb\") pod \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.654515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-config\") pod \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.654547 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-svc\") pod \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.654632 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-665vc\" (UniqueName: \"kubernetes.io/projected/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-kube-api-access-665vc\") pod \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\" (UID: \"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d\") " Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.688358 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-kube-api-access-665vc" (OuterVolumeSpecName: "kube-api-access-665vc") pod "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" (UID: "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d"). InnerVolumeSpecName "kube-api-access-665vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.745801 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" (UID: "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.754336 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" (UID: "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.758798 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-665vc\" (UniqueName: \"kubernetes.io/projected/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-kube-api-access-665vc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.758980 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.759078 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.774662 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-config" (OuterVolumeSpecName: "config") pod "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" (UID: "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.808173 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" (UID: "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.809168 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" (UID: "9b8c483e-d74b-4ae3-8b53-419b8b46fb7d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.860450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" event={"ID":"9b8c483e-d74b-4ae3-8b53-419b8b46fb7d","Type":"ContainerDied","Data":"7adaccee56d76c8a0ed018de9109b89b3c3f7b70549f1e5b3811e72f18fce2f6"} Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.860480 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66898c8f85-lp5qv" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.860503 4945 scope.go:117] "RemoveContainer" containerID="d9cc77780980160bead7c1d8586ff3e120885d1660fbbbb5478dce92676aeb65" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.861590 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.861633 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.861646 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.873828 4945 generic.go:334] "Generic (PLEG): container finished" podID="ce3b16c0-752d-4c25-9565-383b361391de" containerID="ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8" exitCode=0 Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.873894 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8787f869-sll46" event={"ID":"ce3b16c0-752d-4c25-9565-383b361391de","Type":"ContainerDied","Data":"ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8"} Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.881532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d41d3b-2a80-42be-b829-f627d3225331","Type":"ContainerStarted","Data":"d07e269571002d3400eedb751bc4a2dce404f657e054489129cdc69060c1516c"} Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.888244 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 15:35:17 crc kubenswrapper[4945]: E1008 15:35:17.888706 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerName="dnsmasq-dns" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.888720 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerName="dnsmasq-dns" Oct 08 15:35:17 crc kubenswrapper[4945]: E1008 15:35:17.888742 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerName="init" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.888747 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerName="init" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.888934 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" containerName="dnsmasq-dns" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.889651 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.896470 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.896663 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-h7kzj" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.896910 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.907417 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.913184 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66898c8f85-lp5qv"] Oct 08 15:35:17 crc kubenswrapper[4945]: I1008 15:35:17.941156 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66898c8f85-lp5qv"] Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.052795 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b8c483e-d74b-4ae3-8b53-419b8b46fb7d" path="/var/lib/kubelet/pods/9b8c483e-d74b-4ae3-8b53-419b8b46fb7d/volumes" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.073642 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223e741f-c7cc-465f-a4bc-804bfff7a6ef-combined-ca-bundle\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.074017 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/223e741f-c7cc-465f-a4bc-804bfff7a6ef-openstack-config-secret\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.074067 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9ngb\" (UniqueName: \"kubernetes.io/projected/223e741f-c7cc-465f-a4bc-804bfff7a6ef-kube-api-access-z9ngb\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.074258 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/223e741f-c7cc-465f-a4bc-804bfff7a6ef-openstack-config\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.176917 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/223e741f-c7cc-465f-a4bc-804bfff7a6ef-openstack-config-secret\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.176990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9ngb\" (UniqueName: \"kubernetes.io/projected/223e741f-c7cc-465f-a4bc-804bfff7a6ef-kube-api-access-z9ngb\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.177178 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/223e741f-c7cc-465f-a4bc-804bfff7a6ef-openstack-config\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.177261 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223e741f-c7cc-465f-a4bc-804bfff7a6ef-combined-ca-bundle\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.179399 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/223e741f-c7cc-465f-a4bc-804bfff7a6ef-openstack-config\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.182765 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223e741f-c7cc-465f-a4bc-804bfff7a6ef-combined-ca-bundle\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.183128 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/223e741f-c7cc-465f-a4bc-804bfff7a6ef-openstack-config-secret\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.196741 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9ngb\" (UniqueName: \"kubernetes.io/projected/223e741f-c7cc-465f-a4bc-804bfff7a6ef-kube-api-access-z9ngb\") pod \"openstackclient\" (UID: \"223e741f-c7cc-465f-a4bc-804bfff7a6ef\") " pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.239083 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.672810 4945 scope.go:117] "RemoveContainer" containerID="b497f9b12634110643fbf43f63ed784c2a55eb797ab9d0cdd22e91df4c713458" Oct 08 15:35:18 crc kubenswrapper[4945]: I1008 15:35:18.921312 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d41d3b-2a80-42be-b829-f627d3225331","Type":"ContainerStarted","Data":"8539ff391b461261b020a7001064b70c16ef1bf5f9fbd2b60323d83be1892fa5"} Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.057971 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.535024 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.566973 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.943215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"223e741f-c7cc-465f-a4bc-804bfff7a6ef","Type":"ContainerStarted","Data":"b180e5f1ea924b8d3f7b6f758bc1d73b294600cfc9d0d0b92e6d42d7731a6961"} Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.972345 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8787f869-sll46" event={"ID":"ce3b16c0-752d-4c25-9565-383b361391de","Type":"ContainerStarted","Data":"df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9"} Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.972650 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.989157 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api-log" containerID="cri-o://8539ff391b461261b020a7001064b70c16ef1bf5f9fbd2b60323d83be1892fa5" gracePeriod=30 Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.989483 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d41d3b-2a80-42be-b829-f627d3225331","Type":"ContainerStarted","Data":"cb0eaeeb5449edac98afcf1c85c5c37faadd8c2769a7ea49ea2f07bb0b3f43c1"} Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.989530 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 15:35:19 crc kubenswrapper[4945]: I1008 15:35:19.989559 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api" containerID="cri-o://cb0eaeeb5449edac98afcf1c85c5c37faadd8c2769a7ea49ea2f07bb0b3f43c1" gracePeriod=30 Oct 08 15:35:20 crc kubenswrapper[4945]: I1008 15:35:20.013760 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d8787f869-sll46" podStartSLOduration=5.013746 podStartE2EDuration="5.013746s" podCreationTimestamp="2025-10-08 15:35:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:20.013155615 +0000 UTC m=+1329.367070516" watchObservedRunningTime="2025-10-08 15:35:20.013746 +0000 UTC m=+1329.367660901" Oct 08 15:35:20 crc kubenswrapper[4945]: I1008 15:35:20.067097 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.06707826 podStartE2EDuration="5.06707826s" podCreationTimestamp="2025-10-08 15:35:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:20.038343938 +0000 UTC m=+1329.392258849" watchObservedRunningTime="2025-10-08 15:35:20.06707826 +0000 UTC m=+1329.420993161" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.036177 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62c8d55c-c623-4d14-8117-91da2634ad6b","Type":"ContainerStarted","Data":"46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027"} Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.050661 4945 generic.go:334] "Generic (PLEG): container finished" podID="3dec4ad0-549d-49a0-9749-86b21bbaf987" containerID="481ea462b2ffe45c8b6a5bf15ad10dc23616f5609ecafac7d281205ea1afd821" exitCode=0 Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.050713 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ztjf5" event={"ID":"3dec4ad0-549d-49a0-9749-86b21bbaf987","Type":"ContainerDied","Data":"481ea462b2ffe45c8b6a5bf15ad10dc23616f5609ecafac7d281205ea1afd821"} Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.063007 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6d41d3b-2a80-42be-b829-f627d3225331" containerID="cb0eaeeb5449edac98afcf1c85c5c37faadd8c2769a7ea49ea2f07bb0b3f43c1" exitCode=0 Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.063036 4945 generic.go:334] "Generic (PLEG): container finished" podID="b6d41d3b-2a80-42be-b829-f627d3225331" containerID="8539ff391b461261b020a7001064b70c16ef1bf5f9fbd2b60323d83be1892fa5" exitCode=143 Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.063404 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d41d3b-2a80-42be-b829-f627d3225331","Type":"ContainerDied","Data":"cb0eaeeb5449edac98afcf1c85c5c37faadd8c2769a7ea49ea2f07bb0b3f43c1"} Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.063469 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d41d3b-2a80-42be-b829-f627d3225331","Type":"ContainerDied","Data":"8539ff391b461261b020a7001064b70c16ef1bf5f9fbd2b60323d83be1892fa5"} Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.200742 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352595 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d41d3b-2a80-42be-b829-f627d3225331-etc-machine-id\") pod \"b6d41d3b-2a80-42be-b829-f627d3225331\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352641 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-combined-ca-bundle\") pod \"b6d41d3b-2a80-42be-b829-f627d3225331\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352674 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6bcb\" (UniqueName: \"kubernetes.io/projected/b6d41d3b-2a80-42be-b829-f627d3225331-kube-api-access-m6bcb\") pod \"b6d41d3b-2a80-42be-b829-f627d3225331\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352748 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6d41d3b-2a80-42be-b829-f627d3225331-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b6d41d3b-2a80-42be-b829-f627d3225331" (UID: "b6d41d3b-2a80-42be-b829-f627d3225331"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352810 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d41d3b-2a80-42be-b829-f627d3225331-logs\") pod \"b6d41d3b-2a80-42be-b829-f627d3225331\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352878 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data-custom\") pod \"b6d41d3b-2a80-42be-b829-f627d3225331\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352905 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data\") pod \"b6d41d3b-2a80-42be-b829-f627d3225331\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.352947 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-scripts\") pod \"b6d41d3b-2a80-42be-b829-f627d3225331\" (UID: \"b6d41d3b-2a80-42be-b829-f627d3225331\") " Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.353350 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d41d3b-2a80-42be-b829-f627d3225331-logs" (OuterVolumeSpecName: "logs") pod "b6d41d3b-2a80-42be-b829-f627d3225331" (UID: "b6d41d3b-2a80-42be-b829-f627d3225331"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.353795 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d41d3b-2a80-42be-b829-f627d3225331-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.353814 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d41d3b-2a80-42be-b829-f627d3225331-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.358647 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d41d3b-2a80-42be-b829-f627d3225331-kube-api-access-m6bcb" (OuterVolumeSpecName: "kube-api-access-m6bcb") pod "b6d41d3b-2a80-42be-b829-f627d3225331" (UID: "b6d41d3b-2a80-42be-b829-f627d3225331"). InnerVolumeSpecName "kube-api-access-m6bcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.363157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-scripts" (OuterVolumeSpecName: "scripts") pod "b6d41d3b-2a80-42be-b829-f627d3225331" (UID: "b6d41d3b-2a80-42be-b829-f627d3225331"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.396258 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b6d41d3b-2a80-42be-b829-f627d3225331" (UID: "b6d41d3b-2a80-42be-b829-f627d3225331"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.424245 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d41d3b-2a80-42be-b829-f627d3225331" (UID: "b6d41d3b-2a80-42be-b829-f627d3225331"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.451201 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data" (OuterVolumeSpecName: "config-data") pod "b6d41d3b-2a80-42be-b829-f627d3225331" (UID: "b6d41d3b-2a80-42be-b829-f627d3225331"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.455568 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.455777 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6bcb\" (UniqueName: \"kubernetes.io/projected/b6d41d3b-2a80-42be-b829-f627d3225331-kube-api-access-m6bcb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.455863 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.455943 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.456096 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d41d3b-2a80-42be-b829-f627d3225331-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:21 crc kubenswrapper[4945]: I1008 15:35:21.834412 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.002922 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5588d6b8d6-7nmw2" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.078546 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c469d5fc4-4hmvj"] Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.085892 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c469d5fc4-4hmvj" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api-log" containerID="cri-o://4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5" gracePeriod=30 Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.086290 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c469d5fc4-4hmvj" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api" containerID="cri-o://886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc" gracePeriod=30 Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.098942 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62c8d55c-c623-4d14-8117-91da2634ad6b","Type":"ContainerStarted","Data":"2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75"} Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.135398 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.135387 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d41d3b-2a80-42be-b829-f627d3225331","Type":"ContainerDied","Data":"d07e269571002d3400eedb751bc4a2dce404f657e054489129cdc69060c1516c"} Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.137572 4945 scope.go:117] "RemoveContainer" containerID="cb0eaeeb5449edac98afcf1c85c5c37faadd8c2769a7ea49ea2f07bb0b3f43c1" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.155013 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.641760462 podStartE2EDuration="7.154972963s" podCreationTimestamp="2025-10-08 15:35:15 +0000 UTC" firstStartedPulling="2025-10-08 15:35:16.337001831 +0000 UTC m=+1325.690916732" lastFinishedPulling="2025-10-08 15:35:19.850214332 +0000 UTC m=+1329.204129233" observedRunningTime="2025-10-08 15:35:22.140760295 +0000 UTC m=+1331.494675196" watchObservedRunningTime="2025-10-08 15:35:22.154972963 +0000 UTC m=+1331.508887864" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.542263 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.568307 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.572144 4945 scope.go:117] "RemoveContainer" containerID="8539ff391b461261b020a7001064b70c16ef1bf5f9fbd2b60323d83be1892fa5" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.579025 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:22 crc kubenswrapper[4945]: E1008 15:35:22.579583 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.579609 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api" Oct 08 15:35:22 crc kubenswrapper[4945]: E1008 15:35:22.579633 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api-log" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.579641 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api-log" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.579871 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.579888 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" containerName="cinder-api-log" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.590445 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.592185 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.598566 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.598820 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.598995 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf74w\" (UniqueName: \"kubernetes.io/projected/c1bfcac1-b1e7-431f-b145-62a2da972214-kube-api-access-tf74w\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1bfcac1-b1e7-431f-b145-62a2da972214-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693248 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693291 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693310 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-config-data\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693353 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfcac1-b1e7-431f-b145-62a2da972214-logs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693390 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-config-data-custom\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.693415 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-scripts\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.708592 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:35:22 crc kubenswrapper[4945]: E1008 15:35:22.780536 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d41d3b_2a80_42be_b829_f627d3225331.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604863fd_6838_43ef_96a9_393af1ec3c2e.slice/crio-conmon-4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5.scope\": RecentStats: unable to find data in memory cache]" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.794472 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-combined-ca-bundle\") pod \"3dec4ad0-549d-49a0-9749-86b21bbaf987\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.794643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-config\") pod \"3dec4ad0-549d-49a0-9749-86b21bbaf987\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.794817 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsg9v\" (UniqueName: \"kubernetes.io/projected/3dec4ad0-549d-49a0-9749-86b21bbaf987-kube-api-access-tsg9v\") pod \"3dec4ad0-549d-49a0-9749-86b21bbaf987\" (UID: \"3dec4ad0-549d-49a0-9749-86b21bbaf987\") " Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795102 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795134 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-config-data\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795180 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfcac1-b1e7-431f-b145-62a2da972214-logs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-config-data-custom\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-scripts\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf74w\" (UniqueName: \"kubernetes.io/projected/c1bfcac1-b1e7-431f-b145-62a2da972214-kube-api-access-tf74w\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795281 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1bfcac1-b1e7-431f-b145-62a2da972214-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.795367 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1bfcac1-b1e7-431f-b145-62a2da972214-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.807756 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfcac1-b1e7-431f-b145-62a2da972214-logs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.808037 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-scripts\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.808251 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-config-data\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.813376 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.816442 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-config-data-custom\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.825883 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.827420 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dec4ad0-549d-49a0-9749-86b21bbaf987-kube-api-access-tsg9v" (OuterVolumeSpecName: "kube-api-access-tsg9v") pod "3dec4ad0-549d-49a0-9749-86b21bbaf987" (UID: "3dec4ad0-549d-49a0-9749-86b21bbaf987"). InnerVolumeSpecName "kube-api-access-tsg9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.827856 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1bfcac1-b1e7-431f-b145-62a2da972214-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.832273 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf74w\" (UniqueName: \"kubernetes.io/projected/c1bfcac1-b1e7-431f-b145-62a2da972214-kube-api-access-tf74w\") pod \"cinder-api-0\" (UID: \"c1bfcac1-b1e7-431f-b145-62a2da972214\") " pod="openstack/cinder-api-0" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.866445 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-config" (OuterVolumeSpecName: "config") pod "3dec4ad0-549d-49a0-9749-86b21bbaf987" (UID: "3dec4ad0-549d-49a0-9749-86b21bbaf987"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.877557 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dec4ad0-549d-49a0-9749-86b21bbaf987" (UID: "3dec4ad0-549d-49a0-9749-86b21bbaf987"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.898080 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsg9v\" (UniqueName: \"kubernetes.io/projected/3dec4ad0-549d-49a0-9749-86b21bbaf987-kube-api-access-tsg9v\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.898142 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:22 crc kubenswrapper[4945]: I1008 15:35:22.898155 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dec4ad0-549d-49a0-9749-86b21bbaf987-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.019448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.173231 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ztjf5" event={"ID":"3dec4ad0-549d-49a0-9749-86b21bbaf987","Type":"ContainerDied","Data":"c3d7f8797e872ee69b65eef0aeb9ba409af0ff507f794bfc3e812bbeb81bcc8b"} Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.173274 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3d7f8797e872ee69b65eef0aeb9ba409af0ff507f794bfc3e812bbeb81bcc8b" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.173342 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ztjf5" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.191631 4945 generic.go:334] "Generic (PLEG): container finished" podID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerID="4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5" exitCode=143 Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.191817 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c469d5fc4-4hmvj" event={"ID":"604863fd-6838-43ef-96a9-393af1ec3c2e","Type":"ContainerDied","Data":"4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5"} Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.263578 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8787f869-sll46"] Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.263778 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d8787f869-sll46" podUID="ce3b16c0-752d-4c25-9565-383b361391de" containerName="dnsmasq-dns" containerID="cri-o://df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9" gracePeriod=10 Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.292368 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58ccfccdcf-f5tsd"] Oct 08 15:35:23 crc kubenswrapper[4945]: E1008 15:35:23.292817 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dec4ad0-549d-49a0-9749-86b21bbaf987" containerName="neutron-db-sync" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.292838 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dec4ad0-549d-49a0-9749-86b21bbaf987" containerName="neutron-db-sync" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.293032 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dec4ad0-549d-49a0-9749-86b21bbaf987" containerName="neutron-db-sync" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.294010 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.327971 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58ccfccdcf-f5tsd"] Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.415976 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-swift-storage-0\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.416424 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-svc\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.416530 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-sb\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.416558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fpxx\" (UniqueName: \"kubernetes.io/projected/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-kube-api-access-8fpxx\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.416618 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-config\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.417589 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-nb\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.519201 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-sb\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.519259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fpxx\" (UniqueName: \"kubernetes.io/projected/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-kube-api-access-8fpxx\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.519304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-config\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.519347 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-nb\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.519468 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-swift-storage-0\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.519516 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-svc\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.524872 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-sb\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.527015 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-config\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.527268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-nb\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.527338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-swift-storage-0\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.536300 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-779fb94656-nstv6"] Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.537832 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.540574 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-svc\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.543525 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.548661 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-779fb94656-nstv6"] Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.549385 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-45f87" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.549627 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.551102 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fpxx\" (UniqueName: \"kubernetes.io/projected/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-kube-api-access-8fpxx\") pod \"dnsmasq-dns-58ccfccdcf-f5tsd\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.552227 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.625227 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-combined-ca-bundle\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.625344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67fqp\" (UniqueName: \"kubernetes.io/projected/48be9df6-e6b0-4345-8237-ea4a52cd03f7-kube-api-access-67fqp\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.625445 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-config\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.625479 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-ovndb-tls-certs\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.625519 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-httpd-config\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.728975 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-combined-ca-bundle\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.729062 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67fqp\" (UniqueName: \"kubernetes.io/projected/48be9df6-e6b0-4345-8237-ea4a52cd03f7-kube-api-access-67fqp\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.729149 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-config\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.729200 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-ovndb-tls-certs\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.729223 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-httpd-config\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.729943 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.741886 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-combined-ca-bundle\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.748993 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-config\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.753462 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-ovndb-tls-certs\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.754132 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-httpd-config\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.760940 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67fqp\" (UniqueName: \"kubernetes.io/projected/48be9df6-e6b0-4345-8237-ea4a52cd03f7-kube-api-access-67fqp\") pod \"neutron-779fb94656-nstv6\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.772307 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 15:35:23 crc kubenswrapper[4945]: I1008 15:35:23.936763 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.046043 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d41d3b-2a80-42be-b829-f627d3225331" path="/var/lib/kubelet/pods/b6d41d3b-2a80-42be-b829-f627d3225331/volumes" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.243863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1bfcac1-b1e7-431f-b145-62a2da972214","Type":"ContainerStarted","Data":"d72c235ee5692af2ef5b5a9e111d25b5bbfd6695f4c3e1a641e5156b8d0807b4"} Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.249600 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.252715 4945 generic.go:334] "Generic (PLEG): container finished" podID="ce3b16c0-752d-4c25-9565-383b361391de" containerID="df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9" exitCode=0 Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.252758 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8787f869-sll46" event={"ID":"ce3b16c0-752d-4c25-9565-383b361391de","Type":"ContainerDied","Data":"df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9"} Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.252787 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8787f869-sll46" event={"ID":"ce3b16c0-752d-4c25-9565-383b361391de","Type":"ContainerDied","Data":"eba61c560b4954b094d617dc475adb2d764975645768c36c7d671f7c981f7119"} Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.252805 4945 scope.go:117] "RemoveContainer" containerID="df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.356215 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck7xk\" (UniqueName: \"kubernetes.io/projected/ce3b16c0-752d-4c25-9565-383b361391de-kube-api-access-ck7xk\") pod \"ce3b16c0-752d-4c25-9565-383b361391de\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.356289 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-nb\") pod \"ce3b16c0-752d-4c25-9565-383b361391de\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.356320 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-config\") pod \"ce3b16c0-752d-4c25-9565-383b361391de\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.356354 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-swift-storage-0\") pod \"ce3b16c0-752d-4c25-9565-383b361391de\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.356537 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-sb\") pod \"ce3b16c0-752d-4c25-9565-383b361391de\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.356578 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-svc\") pod \"ce3b16c0-752d-4c25-9565-383b361391de\" (UID: \"ce3b16c0-752d-4c25-9565-383b361391de\") " Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.372317 4945 scope.go:117] "RemoveContainer" containerID="ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.377055 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3b16c0-752d-4c25-9565-383b361391de-kube-api-access-ck7xk" (OuterVolumeSpecName: "kube-api-access-ck7xk") pod "ce3b16c0-752d-4c25-9565-383b361391de" (UID: "ce3b16c0-752d-4c25-9565-383b361391de"). InnerVolumeSpecName "kube-api-access-ck7xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.452699 4945 scope.go:117] "RemoveContainer" containerID="df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9" Oct 08 15:35:24 crc kubenswrapper[4945]: E1008 15:35:24.456086 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9\": container with ID starting with df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9 not found: ID does not exist" containerID="df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.456178 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9"} err="failed to get container status \"df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9\": rpc error: code = NotFound desc = could not find container \"df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9\": container with ID starting with df3edb72faa0a866c5eef5b9fa2047d3efb3531dd7634e0853c672ae7e61f5e9 not found: ID does not exist" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.456213 4945 scope.go:117] "RemoveContainer" containerID="ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8" Oct 08 15:35:24 crc kubenswrapper[4945]: E1008 15:35:24.457466 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8\": container with ID starting with ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8 not found: ID does not exist" containerID="ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.457499 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8"} err="failed to get container status \"ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8\": rpc error: code = NotFound desc = could not find container \"ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8\": container with ID starting with ca0081101e6327cc091de7731f1ce9ac232be6af8b2d524e0c103e47ffa71db8 not found: ID does not exist" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.459565 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck7xk\" (UniqueName: \"kubernetes.io/projected/ce3b16c0-752d-4c25-9565-383b361391de-kube-api-access-ck7xk\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.464734 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce3b16c0-752d-4c25-9565-383b361391de" (UID: "ce3b16c0-752d-4c25-9565-383b361391de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.479854 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce3b16c0-752d-4c25-9565-383b361391de" (UID: "ce3b16c0-752d-4c25-9565-383b361391de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.523633 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ce3b16c0-752d-4c25-9565-383b361391de" (UID: "ce3b16c0-752d-4c25-9565-383b361391de"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.537235 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce3b16c0-752d-4c25-9565-383b361391de" (UID: "ce3b16c0-752d-4c25-9565-383b361391de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.540065 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58ccfccdcf-f5tsd"] Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.543255 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-config" (OuterVolumeSpecName: "config") pod "ce3b16c0-752d-4c25-9565-383b361391de" (UID: "ce3b16c0-752d-4c25-9565-383b361391de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.562102 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.562167 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.562180 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.562192 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.562204 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3b16c0-752d-4c25-9565-383b361391de-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:24 crc kubenswrapper[4945]: I1008 15:35:24.768649 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-779fb94656-nstv6"] Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.024527 4945 scope.go:117] "RemoveContainer" containerID="a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6" Oct 08 15:35:25 crc kubenswrapper[4945]: E1008 15:35:25.024979 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.197473 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c469d5fc4-4hmvj" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.170:9311/healthcheck\": read tcp 10.217.0.2:58550->10.217.0.170:9311: read: connection reset by peer" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.197514 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c469d5fc4-4hmvj" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.170:9311/healthcheck\": read tcp 10.217.0.2:58566->10.217.0.170:9311: read: connection reset by peer" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.267206 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8787f869-sll46" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.272831 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-779fb94656-nstv6" event={"ID":"48be9df6-e6b0-4345-8237-ea4a52cd03f7","Type":"ContainerStarted","Data":"43d7c5ffa4eeae66140beeb5081c713ba6729a331de1f70c55ac886e9dff4589"} Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.272869 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-779fb94656-nstv6" event={"ID":"48be9df6-e6b0-4345-8237-ea4a52cd03f7","Type":"ContainerStarted","Data":"d3790c0a371b32b986f41780b0d6e6e4d8e34cf865402dad003dcc6dab224012"} Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.278879 4945 generic.go:334] "Generic (PLEG): container finished" podID="0646f7b0-6abf-436c-976f-59ab119525ad" containerID="d51a29b99688467a7a64d40bba0f998eb8efc47d4f0e03d26b39c3dc69ab162f" exitCode=0 Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.278982 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4bnbb" event={"ID":"0646f7b0-6abf-436c-976f-59ab119525ad","Type":"ContainerDied","Data":"d51a29b99688467a7a64d40bba0f998eb8efc47d4f0e03d26b39c3dc69ab162f"} Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.282361 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1bfcac1-b1e7-431f-b145-62a2da972214","Type":"ContainerStarted","Data":"db6df432a731f5ac56507780836e252f5049eaf02fd2990224de662192a12222"} Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.285214 4945 generic.go:334] "Generic (PLEG): container finished" podID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerID="b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e" exitCode=0 Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.285257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" event={"ID":"7e9e6343-668c-4222-bc2b-86bc8c9f9e99","Type":"ContainerDied","Data":"b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e"} Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.285282 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" event={"ID":"7e9e6343-668c-4222-bc2b-86bc8c9f9e99","Type":"ContainerStarted","Data":"fa13965ac1d65729299dbaaa460829eb734044ddccffc2b4448e7d0e96bd77a4"} Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.350636 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8787f869-sll46"] Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.358219 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d8787f869-sll46"] Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.373937 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.744502 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.887221 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.986241 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604863fd-6838-43ef-96a9-393af1ec3c2e-logs\") pod \"604863fd-6838-43ef-96a9-393af1ec3c2e\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.986285 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ljgl\" (UniqueName: \"kubernetes.io/projected/604863fd-6838-43ef-96a9-393af1ec3c2e-kube-api-access-9ljgl\") pod \"604863fd-6838-43ef-96a9-393af1ec3c2e\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.986361 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data-custom\") pod \"604863fd-6838-43ef-96a9-393af1ec3c2e\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.986438 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data\") pod \"604863fd-6838-43ef-96a9-393af1ec3c2e\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.986553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-combined-ca-bundle\") pod \"604863fd-6838-43ef-96a9-393af1ec3c2e\" (UID: \"604863fd-6838-43ef-96a9-393af1ec3c2e\") " Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.986810 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/604863fd-6838-43ef-96a9-393af1ec3c2e-logs" (OuterVolumeSpecName: "logs") pod "604863fd-6838-43ef-96a9-393af1ec3c2e" (UID: "604863fd-6838-43ef-96a9-393af1ec3c2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:25 crc kubenswrapper[4945]: I1008 15:35:25.986964 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604863fd-6838-43ef-96a9-393af1ec3c2e-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.005511 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/604863fd-6838-43ef-96a9-393af1ec3c2e-kube-api-access-9ljgl" (OuterVolumeSpecName: "kube-api-access-9ljgl") pod "604863fd-6838-43ef-96a9-393af1ec3c2e" (UID: "604863fd-6838-43ef-96a9-393af1ec3c2e"). InnerVolumeSpecName "kube-api-access-9ljgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.009318 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "604863fd-6838-43ef-96a9-393af1ec3c2e" (UID: "604863fd-6838-43ef-96a9-393af1ec3c2e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.023582 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "604863fd-6838-43ef-96a9-393af1ec3c2e" (UID: "604863fd-6838-43ef-96a9-393af1ec3c2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.035998 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3b16c0-752d-4c25-9565-383b361391de" path="/var/lib/kubelet/pods/ce3b16c0-752d-4c25-9565-383b361391de/volumes" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.036711 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data" (OuterVolumeSpecName: "config-data") pod "604863fd-6838-43ef-96a9-393af1ec3c2e" (UID: "604863fd-6838-43ef-96a9-393af1ec3c2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.088440 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ljgl\" (UniqueName: \"kubernetes.io/projected/604863fd-6838-43ef-96a9-393af1ec3c2e-kube-api-access-9ljgl\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.088477 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.088486 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.088495 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604863fd-6838-43ef-96a9-393af1ec3c2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.308119 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1bfcac1-b1e7-431f-b145-62a2da972214","Type":"ContainerStarted","Data":"6c0ff25832a436bbaf43aaa420c0be567c7d6a4472681e0861f9b4be56ea2d46"} Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.309218 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.341065 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" event={"ID":"7e9e6343-668c-4222-bc2b-86bc8c9f9e99","Type":"ContainerStarted","Data":"8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be"} Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.341282 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.351238 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-779fb94656-nstv6" event={"ID":"48be9df6-e6b0-4345-8237-ea4a52cd03f7","Type":"ContainerStarted","Data":"745d9a9caca513603cff2b68e8fe7563b8ed2a7ab5517886916c060ca002e593"} Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.352027 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.357416 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.357405538 podStartE2EDuration="4.357405538s" podCreationTimestamp="2025-10-08 15:35:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:26.355244014 +0000 UTC m=+1335.709158915" watchObservedRunningTime="2025-10-08 15:35:26.357405538 +0000 UTC m=+1335.711320439" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.371867 4945 generic.go:334] "Generic (PLEG): container finished" podID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerID="886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc" exitCode=0 Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.372064 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c469d5fc4-4hmvj" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.372181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c469d5fc4-4hmvj" event={"ID":"604863fd-6838-43ef-96a9-393af1ec3c2e","Type":"ContainerDied","Data":"886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc"} Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.372244 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c469d5fc4-4hmvj" event={"ID":"604863fd-6838-43ef-96a9-393af1ec3c2e","Type":"ContainerDied","Data":"78847e284c5fb3ad8501f7810c75e5c23810856820ca8c8eaa83ffa8d067ceae"} Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.372268 4945 scope.go:117] "RemoveContainer" containerID="886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.420986 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-779fb94656-nstv6" podStartSLOduration=3.420971094 podStartE2EDuration="3.420971094s" podCreationTimestamp="2025-10-08 15:35:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:26.387790521 +0000 UTC m=+1335.741705422" watchObservedRunningTime="2025-10-08 15:35:26.420971094 +0000 UTC m=+1335.774885995" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.422387 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" podStartSLOduration=3.42238159 podStartE2EDuration="3.42238159s" podCreationTimestamp="2025-10-08 15:35:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:26.417498397 +0000 UTC m=+1335.771413298" watchObservedRunningTime="2025-10-08 15:35:26.42238159 +0000 UTC m=+1335.776296491" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.516325 4945 scope.go:117] "RemoveContainer" containerID="4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.539465 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.554498 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c469d5fc4-4hmvj"] Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.557313 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-c469d5fc4-4hmvj"] Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.569277 4945 scope.go:117] "RemoveContainer" containerID="886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc" Oct 08 15:35:26 crc kubenswrapper[4945]: E1008 15:35:26.570931 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc\": container with ID starting with 886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc not found: ID does not exist" containerID="886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.570970 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc"} err="failed to get container status \"886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc\": rpc error: code = NotFound desc = could not find container \"886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc\": container with ID starting with 886f6d2a2f0f632ba7c73894b5806cc29d2144ff1e9fb1b328b5b8b2af62f2fc not found: ID does not exist" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.570991 4945 scope.go:117] "RemoveContainer" containerID="4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5" Oct 08 15:35:26 crc kubenswrapper[4945]: E1008 15:35:26.571874 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5\": container with ID starting with 4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5 not found: ID does not exist" containerID="4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.571904 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5"} err="failed to get container status \"4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5\": rpc error: code = NotFound desc = could not find container \"4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5\": container with ID starting with 4deaa38b384cb81da28d1ada6216e131e845a126af31291f9a71925ff59920e5 not found: ID does not exist" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.679635 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5988468485-ctpxh"] Oct 08 15:35:26 crc kubenswrapper[4945]: E1008 15:35:26.680014 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3b16c0-752d-4c25-9565-383b361391de" containerName="init" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.680029 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3b16c0-752d-4c25-9565-383b361391de" containerName="init" Oct 08 15:35:26 crc kubenswrapper[4945]: E1008 15:35:26.680040 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3b16c0-752d-4c25-9565-383b361391de" containerName="dnsmasq-dns" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.680046 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3b16c0-752d-4c25-9565-383b361391de" containerName="dnsmasq-dns" Oct 08 15:35:26 crc kubenswrapper[4945]: E1008 15:35:26.680068 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api-log" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.680075 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api-log" Oct 08 15:35:26 crc kubenswrapper[4945]: E1008 15:35:26.680088 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.680093 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.680281 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3b16c0-752d-4c25-9565-383b361391de" containerName="dnsmasq-dns" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.680294 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.680313 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" containerName="barbican-api-log" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.681252 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.684593 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.684704 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.706320 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlfqv\" (UniqueName: \"kubernetes.io/projected/acfb96c2-ab3d-4e3f-86ce-c6befd371617-kube-api-access-xlfqv\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.706411 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-ovndb-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.706441 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-combined-ca-bundle\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.706461 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-internal-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.706562 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-public-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.706596 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-httpd-config\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.706635 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-config\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.717964 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5988468485-ctpxh"] Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.808662 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-httpd-config\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.808715 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-config\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.808756 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlfqv\" (UniqueName: \"kubernetes.io/projected/acfb96c2-ab3d-4e3f-86ce-c6befd371617-kube-api-access-xlfqv\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.808796 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-ovndb-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.808813 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-combined-ca-bundle\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.808826 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-internal-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.808914 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-public-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.827194 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-config\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.827774 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-internal-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.837084 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-httpd-config\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.837859 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-public-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.843500 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-ovndb-tls-certs\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.844496 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfb96c2-ab3d-4e3f-86ce-c6befd371617-combined-ca-bundle\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.852397 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlfqv\" (UniqueName: \"kubernetes.io/projected/acfb96c2-ab3d-4e3f-86ce-c6befd371617-kube-api-access-xlfqv\") pod \"neutron-5988468485-ctpxh\" (UID: \"acfb96c2-ab3d-4e3f-86ce-c6befd371617\") " pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.941058 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.941888 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-central-agent" containerID="cri-o://32502233dfa60024f099b1e52bd8769c232d782ad1560a3c56753539d5a5f6f7" gracePeriod=30 Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.942005 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="proxy-httpd" containerID="cri-o://2f5f9d1ad32e9a7ee9cfa649f015b3c74f56800032144d01bcedd0537b6c0114" gracePeriod=30 Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.942049 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="sg-core" containerID="cri-o://279951b5df62c3b8c95708382849b0f083ceb676b2e74662684f60764ce3c250" gracePeriod=30 Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.942078 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-notification-agent" containerID="cri-o://8395ee59d8bef1a2e368ede09a30682e3b96e97a08837fe250471dad4cb35de5" gracePeriod=30 Oct 08 15:35:26 crc kubenswrapper[4945]: I1008 15:35:26.998957 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.062636 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": read tcp 10.217.0.2:32888->10.217.0.165:3000: read: connection reset by peer" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.135269 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4bnbb" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.164659 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58798f7484-jkqnb" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.166364 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7cfddd55fc-6xk6v"] Oct 08 15:35:27 crc kubenswrapper[4945]: E1008 15:35:27.166834 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" containerName="glance-db-sync" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.166852 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" containerName="glance-db-sync" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.167086 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" containerName="glance-db-sync" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.168089 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.171703 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.171920 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.172041 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.190603 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7cfddd55fc-6xk6v"] Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.233684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-db-sync-config-data\") pod \"0646f7b0-6abf-436c-976f-59ab119525ad\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.233749 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr722\" (UniqueName: \"kubernetes.io/projected/0646f7b0-6abf-436c-976f-59ab119525ad-kube-api-access-qr722\") pod \"0646f7b0-6abf-436c-976f-59ab119525ad\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.233780 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-config-data\") pod \"0646f7b0-6abf-436c-976f-59ab119525ad\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.233844 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-combined-ca-bundle\") pod \"0646f7b0-6abf-436c-976f-59ab119525ad\" (UID: \"0646f7b0-6abf-436c-976f-59ab119525ad\") " Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234183 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-internal-tls-certs\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234226 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-etc-swift\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234351 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-config-data\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234439 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2mdl\" (UniqueName: \"kubernetes.io/projected/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-kube-api-access-w2mdl\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234481 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-public-tls-certs\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234519 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-run-httpd\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234565 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-combined-ca-bundle\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.234602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-log-httpd\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.239354 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0646f7b0-6abf-436c-976f-59ab119525ad-kube-api-access-qr722" (OuterVolumeSpecName: "kube-api-access-qr722") pod "0646f7b0-6abf-436c-976f-59ab119525ad" (UID: "0646f7b0-6abf-436c-976f-59ab119525ad"). InnerVolumeSpecName "kube-api-access-qr722". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.239595 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0646f7b0-6abf-436c-976f-59ab119525ad" (UID: "0646f7b0-6abf-436c-976f-59ab119525ad"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.266577 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0646f7b0-6abf-436c-976f-59ab119525ad" (UID: "0646f7b0-6abf-436c-976f-59ab119525ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.305599 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-config-data" (OuterVolumeSpecName: "config-data") pod "0646f7b0-6abf-436c-976f-59ab119525ad" (UID: "0646f7b0-6abf-436c-976f-59ab119525ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-internal-tls-certs\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-etc-swift\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-config-data\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336344 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2mdl\" (UniqueName: \"kubernetes.io/projected/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-kube-api-access-w2mdl\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336367 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-public-tls-certs\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336392 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-run-httpd\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336417 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-combined-ca-bundle\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336437 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-log-httpd\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336498 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr722\" (UniqueName: \"kubernetes.io/projected/0646f7b0-6abf-436c-976f-59ab119525ad-kube-api-access-qr722\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336508 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336517 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336526 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0646f7b0-6abf-436c-976f-59ab119525ad-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.336955 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-log-httpd\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.340162 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-run-httpd\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.343740 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-etc-swift\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.343959 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-public-tls-certs\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.345019 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-config-data\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.346769 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-combined-ca-bundle\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.348441 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-internal-tls-certs\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.357791 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2mdl\" (UniqueName: \"kubernetes.io/projected/2e5af0dd-db9f-40a7-941c-1c08c0bc5be4-kube-api-access-w2mdl\") pod \"swift-proxy-7cfddd55fc-6xk6v\" (UID: \"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4\") " pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.414780 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4bnbb" event={"ID":"0646f7b0-6abf-436c-976f-59ab119525ad","Type":"ContainerDied","Data":"e70453d1d42cfd4f9c54d651399dd52f250f583dd4fcc0b0420deea266708cc6"} Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.415030 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e70453d1d42cfd4f9c54d651399dd52f250f583dd4fcc0b0420deea266708cc6" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.415128 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4bnbb" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.440160 4945 generic.go:334] "Generic (PLEG): container finished" podID="b00bb163-6180-4acb-a485-095a289db80a" containerID="2f5f9d1ad32e9a7ee9cfa649f015b3c74f56800032144d01bcedd0537b6c0114" exitCode=0 Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.440198 4945 generic.go:334] "Generic (PLEG): container finished" podID="b00bb163-6180-4acb-a485-095a289db80a" containerID="279951b5df62c3b8c95708382849b0f083ceb676b2e74662684f60764ce3c250" exitCode=2 Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.440873 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerDied","Data":"2f5f9d1ad32e9a7ee9cfa649f015b3c74f56800032144d01bcedd0537b6c0114"} Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.440940 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerDied","Data":"279951b5df62c3b8c95708382849b0f083ceb676b2e74662684f60764ce3c250"} Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.441054 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="cinder-scheduler" containerID="cri-o://46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027" gracePeriod=30 Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.441166 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="probe" containerID="cri-o://2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75" gracePeriod=30 Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.534706 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.697624 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58ccfccdcf-f5tsd"] Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.722125 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b84564489-l4jsm"] Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.723677 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.741157 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b84564489-l4jsm"] Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.788481 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5988468485-ctpxh"] Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.848135 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxw6n\" (UniqueName: \"kubernetes.io/projected/c801f135-2df2-4cc1-9f62-dba3929cf17e-kube-api-access-gxw6n\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.848230 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-config\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.848335 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-sb\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.848428 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-svc\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.848569 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-swift-storage-0\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.848799 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-nb\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.949980 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-swift-storage-0\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.950078 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-nb\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.950148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxw6n\" (UniqueName: \"kubernetes.io/projected/c801f135-2df2-4cc1-9f62-dba3929cf17e-kube-api-access-gxw6n\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.950169 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-config\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.950207 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-sb\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.950237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-svc\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.950967 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-svc\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.952678 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-nb\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.953390 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-config\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.954701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-sb\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.956371 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-swift-storage-0\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:27 crc kubenswrapper[4945]: I1008 15:35:27.975136 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxw6n\" (UniqueName: \"kubernetes.io/projected/c801f135-2df2-4cc1-9f62-dba3929cf17e-kube-api-access-gxw6n\") pod \"dnsmasq-dns-5b84564489-l4jsm\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.047632 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="604863fd-6838-43ef-96a9-393af1ec3c2e" path="/var/lib/kubelet/pods/604863fd-6838-43ef-96a9-393af1ec3c2e/volumes" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.068598 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.335671 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7cfddd55fc-6xk6v"] Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.482893 4945 generic.go:334] "Generic (PLEG): container finished" podID="b00bb163-6180-4acb-a485-095a289db80a" containerID="32502233dfa60024f099b1e52bd8769c232d782ad1560a3c56753539d5a5f6f7" exitCode=0 Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.483625 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerDied","Data":"32502233dfa60024f099b1e52bd8769c232d782ad1560a3c56753539d5a5f6f7"} Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.502028 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5988468485-ctpxh" event={"ID":"acfb96c2-ab3d-4e3f-86ce-c6befd371617","Type":"ContainerStarted","Data":"4f6599ef76251baf47ebbb9e5caf31215466f62435ccaadde46d8313c8be3521"} Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.502082 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5988468485-ctpxh" event={"ID":"acfb96c2-ab3d-4e3f-86ce-c6befd371617","Type":"ContainerStarted","Data":"69001c83f7000bf08694c1479840ae93d351270eb7391780f16ff6b45c7de00e"} Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.503119 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.517814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" event={"ID":"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4","Type":"ContainerStarted","Data":"1dfa73d6a5a002e0c8eea4040984ecc0a1dea58a4e1b11626c32d27d2a150075"} Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.518595 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" podUID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerName="dnsmasq-dns" containerID="cri-o://8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be" gracePeriod=10 Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.596870 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5988468485-ctpxh" podStartSLOduration=2.596848028 podStartE2EDuration="2.596848028s" podCreationTimestamp="2025-10-08 15:35:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:28.540195814 +0000 UTC m=+1337.894110725" watchObservedRunningTime="2025-10-08 15:35:28.596848028 +0000 UTC m=+1337.950762929" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.674568 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.700402 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.703773 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.704219 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.704370 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g97m6" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.718936 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.747085 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b84564489-l4jsm"] Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.801220 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.801311 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.801335 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-logs\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.801413 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.801452 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-scripts\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.801468 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xbjr\" (UniqueName: \"kubernetes.io/projected/2bd5462d-d260-44d3-ab01-765e4fe34744-kube-api-access-7xbjr\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.801484 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-config-data\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.905661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.905716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-logs\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.905812 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.905858 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-scripts\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.905877 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xbjr\" (UniqueName: \"kubernetes.io/projected/2bd5462d-d260-44d3-ab01-765e4fe34744-kube-api-access-7xbjr\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.905898 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-config-data\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.905933 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.906334 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.907074 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-logs\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.909610 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.911123 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.911662 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.911947 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.917344 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-scripts\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.920198 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.920830 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.924393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-config-data\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.949439 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:28 crc kubenswrapper[4945]: I1008 15:35:28.954158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xbjr\" (UniqueName: \"kubernetes.io/projected/2bd5462d-d260-44d3-ab01-765e4fe34744-kube-api-access-7xbjr\") pod \"glance-default-external-api-0\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.008617 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w9hm\" (UniqueName: \"kubernetes.io/projected/6f3a5305-f80b-4798-9444-90152292e039-kube-api-access-8w9hm\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.008682 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.008704 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.008773 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.008795 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.008817 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-logs\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.008843 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w9hm\" (UniqueName: \"kubernetes.io/projected/6f3a5305-f80b-4798-9444-90152292e039-kube-api-access-8w9hm\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112342 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112461 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112504 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112552 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-logs\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112603 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.112837 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.113508 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.115613 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.116826 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-logs\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.125472 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.134669 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.143175 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w9hm\" (UniqueName: \"kubernetes.io/projected/6f3a5305-f80b-4798-9444-90152292e039-kube-api-access-8w9hm\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.143472 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.170355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.229669 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.364457 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.442952 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-swift-storage-0\") pod \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.443045 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-svc\") pod \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.443182 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-sb\") pod \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.443333 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-nb\") pod \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.443410 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fpxx\" (UniqueName: \"kubernetes.io/projected/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-kube-api-access-8fpxx\") pod \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.443434 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-config\") pod \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\" (UID: \"7e9e6343-668c-4222-bc2b-86bc8c9f9e99\") " Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.466172 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-kube-api-access-8fpxx" (OuterVolumeSpecName: "kube-api-access-8fpxx") pod "7e9e6343-668c-4222-bc2b-86bc8c9f9e99" (UID: "7e9e6343-668c-4222-bc2b-86bc8c9f9e99"). InnerVolumeSpecName "kube-api-access-8fpxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.559449 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fpxx\" (UniqueName: \"kubernetes.io/projected/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-kube-api-access-8fpxx\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.588788 4945 generic.go:334] "Generic (PLEG): container finished" podID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerID="8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be" exitCode=0 Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.589428 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.589447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" event={"ID":"7e9e6343-668c-4222-bc2b-86bc8c9f9e99","Type":"ContainerDied","Data":"8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.590777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58ccfccdcf-f5tsd" event={"ID":"7e9e6343-668c-4222-bc2b-86bc8c9f9e99","Type":"ContainerDied","Data":"fa13965ac1d65729299dbaaa460829eb734044ddccffc2b4448e7d0e96bd77a4"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.590803 4945 scope.go:117] "RemoveContainer" containerID="8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.595881 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e9e6343-668c-4222-bc2b-86bc8c9f9e99" (UID: "7e9e6343-668c-4222-bc2b-86bc8c9f9e99"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.603353 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e9e6343-668c-4222-bc2b-86bc8c9f9e99" (UID: "7e9e6343-668c-4222-bc2b-86bc8c9f9e99"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.603743 4945 generic.go:334] "Generic (PLEG): container finished" podID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerID="1f429ec667ec642585a50f8a5121fe7c9e1361690f688476e8976d114b401757" exitCode=0 Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.603831 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" event={"ID":"c801f135-2df2-4cc1-9f62-dba3929cf17e","Type":"ContainerDied","Data":"1f429ec667ec642585a50f8a5121fe7c9e1361690f688476e8976d114b401757"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.603863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" event={"ID":"c801f135-2df2-4cc1-9f62-dba3929cf17e","Type":"ContainerStarted","Data":"534672ab02f381e7dbd1cd260ec7a5a53a521aa86707a8445a8bbec8defa83fc"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.604733 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.609920 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e9e6343-668c-4222-bc2b-86bc8c9f9e99" (UID: "7e9e6343-668c-4222-bc2b-86bc8c9f9e99"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:29 crc kubenswrapper[4945]: W1008 15:35:29.617018 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bd5462d_d260_44d3_ab01_765e4fe34744.slice/crio-f63d4e19d792cb0374f3e01b042a3eebd4fb63845c6d6d6e33fb043c3748eeca WatchSource:0}: Error finding container f63d4e19d792cb0374f3e01b042a3eebd4fb63845c6d6d6e33fb043c3748eeca: Status 404 returned error can't find the container with id f63d4e19d792cb0374f3e01b042a3eebd4fb63845c6d6d6e33fb043c3748eeca Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.633555 4945 generic.go:334] "Generic (PLEG): container finished" podID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerID="2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75" exitCode=0 Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.633711 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62c8d55c-c623-4d14-8117-91da2634ad6b","Type":"ContainerDied","Data":"2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.659226 4945 generic.go:334] "Generic (PLEG): container finished" podID="b00bb163-6180-4acb-a485-095a289db80a" containerID="8395ee59d8bef1a2e368ede09a30682e3b96e97a08837fe250471dad4cb35de5" exitCode=0 Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.659281 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerDied","Data":"8395ee59d8bef1a2e368ede09a30682e3b96e97a08837fe250471dad4cb35de5"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.661544 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.661562 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.661573 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.667909 4945 scope.go:117] "RemoveContainer" containerID="b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.676222 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5988468485-ctpxh" event={"ID":"acfb96c2-ab3d-4e3f-86ce-c6befd371617","Type":"ContainerStarted","Data":"dad1af0b65258e0ee745e8cbb7b96fa95e93a4f637db6fce135e8429a069a5fb"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.682992 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-config" (OuterVolumeSpecName: "config") pod "7e9e6343-668c-4222-bc2b-86bc8c9f9e99" (UID: "7e9e6343-668c-4222-bc2b-86bc8c9f9e99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.693697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" event={"ID":"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4","Type":"ContainerStarted","Data":"aad850466f0005c5822d69ee238f0854ddb27d9ec639bcb18e841e26006fabe6"} Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.721680 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e9e6343-668c-4222-bc2b-86bc8c9f9e99" (UID: "7e9e6343-668c-4222-bc2b-86bc8c9f9e99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.763589 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.763620 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9e6343-668c-4222-bc2b-86bc8c9f9e99-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.827446 4945 scope.go:117] "RemoveContainer" containerID="8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be" Oct 08 15:35:29 crc kubenswrapper[4945]: E1008 15:35:29.828127 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be\": container with ID starting with 8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be not found: ID does not exist" containerID="8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.828186 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be"} err="failed to get container status \"8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be\": rpc error: code = NotFound desc = could not find container \"8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be\": container with ID starting with 8e5269448b18f8f9ec116af446355c11eef4966174885c74daa04831d38169be not found: ID does not exist" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.828212 4945 scope.go:117] "RemoveContainer" containerID="b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e" Oct 08 15:35:29 crc kubenswrapper[4945]: E1008 15:35:29.851347 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e\": container with ID starting with b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e not found: ID does not exist" containerID="b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.851684 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e"} err="failed to get container status \"b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e\": rpc error: code = NotFound desc = could not find container \"b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e\": container with ID starting with b9654b3f71540219f067de1306c5fcfe5f8ffdbb8c74f49f48d9fe6a6289370e not found: ID does not exist" Oct 08 15:35:29 crc kubenswrapper[4945]: I1008 15:35:29.997547 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58ccfccdcf-f5tsd"] Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.017696 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58ccfccdcf-f5tsd"] Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.049309 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" path="/var/lib/kubelet/pods/7e9e6343-668c-4222-bc2b-86bc8c9f9e99/volumes" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.050298 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:30 crc kubenswrapper[4945]: W1008 15:35:30.054760 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f3a5305_f80b_4798_9444_90152292e039.slice/crio-ec68f09d4c01517e4795689d35bc937538db86efdf08cedf06a26d50ce9d1c17 WatchSource:0}: Error finding container ec68f09d4c01517e4795689d35bc937538db86efdf08cedf06a26d50ce9d1c17: Status 404 returned error can't find the container with id ec68f09d4c01517e4795689d35bc937538db86efdf08cedf06a26d50ce9d1c17 Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.238863 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.382268 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-log-httpd\") pod \"b00bb163-6180-4acb-a485-095a289db80a\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.382603 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-combined-ca-bundle\") pod \"b00bb163-6180-4acb-a485-095a289db80a\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.382638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4jgh\" (UniqueName: \"kubernetes.io/projected/b00bb163-6180-4acb-a485-095a289db80a-kube-api-access-g4jgh\") pod \"b00bb163-6180-4acb-a485-095a289db80a\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.382688 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-scripts\") pod \"b00bb163-6180-4acb-a485-095a289db80a\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.382711 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-config-data\") pod \"b00bb163-6180-4acb-a485-095a289db80a\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.382775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-run-httpd\") pod \"b00bb163-6180-4acb-a485-095a289db80a\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.382826 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-sg-core-conf-yaml\") pod \"b00bb163-6180-4acb-a485-095a289db80a\" (UID: \"b00bb163-6180-4acb-a485-095a289db80a\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.384232 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b00bb163-6180-4acb-a485-095a289db80a" (UID: "b00bb163-6180-4acb-a485-095a289db80a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.386602 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b00bb163-6180-4acb-a485-095a289db80a" (UID: "b00bb163-6180-4acb-a485-095a289db80a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.393021 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00bb163-6180-4acb-a485-095a289db80a-kube-api-access-g4jgh" (OuterVolumeSpecName: "kube-api-access-g4jgh") pod "b00bb163-6180-4acb-a485-095a289db80a" (UID: "b00bb163-6180-4acb-a485-095a289db80a"). InnerVolumeSpecName "kube-api-access-g4jgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.394555 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-scripts" (OuterVolumeSpecName: "scripts") pod "b00bb163-6180-4acb-a485-095a289db80a" (UID: "b00bb163-6180-4acb-a485-095a289db80a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.460252 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b00bb163-6180-4acb-a485-095a289db80a" (UID: "b00bb163-6180-4acb-a485-095a289db80a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.487486 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.487518 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4jgh\" (UniqueName: \"kubernetes.io/projected/b00bb163-6180-4acb-a485-095a289db80a-kube-api-access-g4jgh\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.487529 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.487537 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00bb163-6180-4acb-a485-095a289db80a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.487545 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.536083 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b00bb163-6180-4acb-a485-095a289db80a" (UID: "b00bb163-6180-4acb-a485-095a289db80a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.591635 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.657057 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-config-data" (OuterVolumeSpecName: "config-data") pod "b00bb163-6180-4acb-a485-095a289db80a" (UID: "b00bb163-6180-4acb-a485-095a289db80a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.672352 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.693814 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb163-6180-4acb-a485-095a289db80a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.725172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" event={"ID":"c801f135-2df2-4cc1-9f62-dba3929cf17e","Type":"ContainerStarted","Data":"59f37be1ffe2d24e9b8c37cf9889ecea154b870df48daa0e792ddd1dd5c01383"} Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.725251 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.761775 4945 generic.go:334] "Generic (PLEG): container finished" podID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerID="46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027" exitCode=0 Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.761835 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62c8d55c-c623-4d14-8117-91da2634ad6b","Type":"ContainerDied","Data":"46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027"} Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.761862 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62c8d55c-c623-4d14-8117-91da2634ad6b","Type":"ContainerDied","Data":"26158ba8609095114810fd0d2879e0e56a5115c958352b0ccf3d0aa5c49d29c4"} Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.761878 4945 scope.go:117] "RemoveContainer" containerID="2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.761989 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.766909 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" podStartSLOduration=3.766883234 podStartE2EDuration="3.766883234s" podCreationTimestamp="2025-10-08 15:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:30.758518214 +0000 UTC m=+1340.112433115" watchObservedRunningTime="2025-10-08 15:35:30.766883234 +0000 UTC m=+1340.120798145" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.779284 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bd5462d-d260-44d3-ab01-765e4fe34744","Type":"ContainerStarted","Data":"f63d4e19d792cb0374f3e01b042a3eebd4fb63845c6d6d6e33fb043c3748eeca"} Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.785517 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00bb163-6180-4acb-a485-095a289db80a","Type":"ContainerDied","Data":"abdb5ca063d09e7861b68b34c9a7c282768c0f44d34e5be13bf843feb873c1e7"} Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.785631 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.787671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6f3a5305-f80b-4798-9444-90152292e039","Type":"ContainerStarted","Data":"ec68f09d4c01517e4795689d35bc937538db86efdf08cedf06a26d50ce9d1c17"} Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.792307 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" event={"ID":"2e5af0dd-db9f-40a7-941c-1c08c0bc5be4","Type":"ContainerStarted","Data":"0941b9071cfd85d119ed63062813936544a1d74f18aa08d4fc8a2fc6b9ac97b9"} Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.795157 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.795192 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.799776 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data\") pod \"62c8d55c-c623-4d14-8117-91da2634ad6b\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.802359 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-combined-ca-bundle\") pod \"62c8d55c-c623-4d14-8117-91da2634ad6b\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.802442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ljdz\" (UniqueName: \"kubernetes.io/projected/62c8d55c-c623-4d14-8117-91da2634ad6b-kube-api-access-8ljdz\") pod \"62c8d55c-c623-4d14-8117-91da2634ad6b\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.802561 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62c8d55c-c623-4d14-8117-91da2634ad6b-etc-machine-id\") pod \"62c8d55c-c623-4d14-8117-91da2634ad6b\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.802620 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data-custom\") pod \"62c8d55c-c623-4d14-8117-91da2634ad6b\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.802668 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-scripts\") pod \"62c8d55c-c623-4d14-8117-91da2634ad6b\" (UID: \"62c8d55c-c623-4d14-8117-91da2634ad6b\") " Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.803696 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c8d55c-c623-4d14-8117-91da2634ad6b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "62c8d55c-c623-4d14-8117-91da2634ad6b" (UID: "62c8d55c-c623-4d14-8117-91da2634ad6b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.820523 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "62c8d55c-c623-4d14-8117-91da2634ad6b" (UID: "62c8d55c-c623-4d14-8117-91da2634ad6b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.830291 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c8d55c-c623-4d14-8117-91da2634ad6b-kube-api-access-8ljdz" (OuterVolumeSpecName: "kube-api-access-8ljdz") pod "62c8d55c-c623-4d14-8117-91da2634ad6b" (UID: "62c8d55c-c623-4d14-8117-91da2634ad6b"). InnerVolumeSpecName "kube-api-access-8ljdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.835694 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-scripts" (OuterVolumeSpecName: "scripts") pod "62c8d55c-c623-4d14-8117-91da2634ad6b" (UID: "62c8d55c-c623-4d14-8117-91da2634ad6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.836066 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" podStartSLOduration=3.836051382 podStartE2EDuration="3.836051382s" podCreationTimestamp="2025-10-08 15:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:30.826199425 +0000 UTC m=+1340.180114316" watchObservedRunningTime="2025-10-08 15:35:30.836051382 +0000 UTC m=+1340.189966283" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.906989 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62c8d55c-c623-4d14-8117-91da2634ad6b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.907773 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.907910 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.907992 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ljdz\" (UniqueName: \"kubernetes.io/projected/62c8d55c-c623-4d14-8117-91da2634ad6b-kube-api-access-8ljdz\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:30 crc kubenswrapper[4945]: I1008 15:35:30.997447 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62c8d55c-c623-4d14-8117-91da2634ad6b" (UID: "62c8d55c-c623-4d14-8117-91da2634ad6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.010492 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.032480 4945 scope.go:117] "RemoveContainer" containerID="46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.055344 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data" (OuterVolumeSpecName: "config-data") pod "62c8d55c-c623-4d14-8117-91da2634ad6b" (UID: "62c8d55c-c623-4d14-8117-91da2634ad6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.115193 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c8d55c-c623-4d14-8117-91da2634ad6b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.189543 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.200290 4945 scope.go:117] "RemoveContainer" containerID="2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.201656 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.208294 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75\": container with ID starting with 2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75 not found: ID does not exist" containerID="2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.208331 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75"} err="failed to get container status \"2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75\": rpc error: code = NotFound desc = could not find container \"2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75\": container with ID starting with 2d8e96bc2d8ce940c47458031a3829b7f99e61a74e633375d5f240cf50bbeb75 not found: ID does not exist" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.208356 4945 scope.go:117] "RemoveContainer" containerID="46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.213950 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027\": container with ID starting with 46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027 not found: ID does not exist" containerID="46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.213984 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027"} err="failed to get container status \"46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027\": rpc error: code = NotFound desc = could not find container \"46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027\": container with ID starting with 46d09a90638ade2fcf48b3d0dc385e1a68d89f4757ae7527f9cc405558d02027 not found: ID does not exist" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.214011 4945 scope.go:117] "RemoveContainer" containerID="2f5f9d1ad32e9a7ee9cfa649f015b3c74f56800032144d01bcedd0537b6c0114" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.231243 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.243625 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244028 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="cinder-scheduler" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244041 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="cinder-scheduler" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244055 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerName="dnsmasq-dns" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244061 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerName="dnsmasq-dns" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244071 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-central-agent" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244079 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-central-agent" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244087 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-notification-agent" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244094 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-notification-agent" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244118 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="sg-core" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244123 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="sg-core" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244143 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="proxy-httpd" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244149 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="proxy-httpd" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244166 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="probe" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244171 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="probe" Oct 08 15:35:31 crc kubenswrapper[4945]: E1008 15:35:31.244185 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerName="init" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244190 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerName="init" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244354 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="proxy-httpd" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244366 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9e6343-668c-4222-bc2b-86bc8c9f9e99" containerName="dnsmasq-dns" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244417 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="cinder-scheduler" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244427 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="sg-core" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244435 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-notification-agent" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244443 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00bb163-6180-4acb-a485-095a289db80a" containerName="ceilometer-central-agent" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.244455 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" containerName="probe" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.245433 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.250442 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.257165 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.267170 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.283989 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.286266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.288048 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.288437 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.293022 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.299393 4945 scope.go:117] "RemoveContainer" containerID="279951b5df62c3b8c95708382849b0f083ceb676b2e74662684f60764ce3c250" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.321686 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.321959 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-scripts\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.321997 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-config-data\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.322029 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vhtl\" (UniqueName: \"kubernetes.io/projected/815a3456-3316-4348-a417-f274fda65c41-kube-api-access-7vhtl\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.322059 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.322126 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/815a3456-3316-4348-a417-f274fda65c41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.376071 4945 scope.go:117] "RemoveContainer" containerID="8395ee59d8bef1a2e368ede09a30682e3b96e97a08837fe250471dad4cb35de5" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424314 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-run-httpd\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424373 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pdrp\" (UniqueName: \"kubernetes.io/projected/e10c8660-ea2f-483c-afbe-7f7e6287968c-kube-api-access-9pdrp\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424408 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/815a3456-3316-4348-a417-f274fda65c41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424427 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-scripts\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424514 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424542 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-log-httpd\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424582 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-scripts\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424621 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424658 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-config-data\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.424708 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vhtl\" (UniqueName: \"kubernetes.io/projected/815a3456-3316-4348-a417-f274fda65c41-kube-api-access-7vhtl\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.425089 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/815a3456-3316-4348-a417-f274fda65c41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.439021 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.441345 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-scripts\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.449177 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-config-data\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.457541 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vhtl\" (UniqueName: \"kubernetes.io/projected/815a3456-3316-4348-a417-f274fda65c41-kube-api-access-7vhtl\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.462648 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/815a3456-3316-4348-a417-f274fda65c41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"815a3456-3316-4348-a417-f274fda65c41\") " pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.529609 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-scripts\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.529705 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-log-httpd\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.529774 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.529799 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.529904 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-run-httpd\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.529930 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pdrp\" (UniqueName: \"kubernetes.io/projected/e10c8660-ea2f-483c-afbe-7f7e6287968c-kube-api-access-9pdrp\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.529966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.535444 4945 scope.go:117] "RemoveContainer" containerID="32502233dfa60024f099b1e52bd8769c232d782ad1560a3c56753539d5a5f6f7" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.538535 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.538879 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-log-httpd\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.540036 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-run-httpd\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.542217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.542868 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-scripts\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.545918 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.574848 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pdrp\" (UniqueName: \"kubernetes.io/projected/e10c8660-ea2f-483c-afbe-7f7e6287968c-kube-api-access-9pdrp\") pod \"ceilometer-0\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.593975 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.631716 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.673513 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.733316 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm2ft\" (UniqueName: \"kubernetes.io/projected/4e99e039-6910-4bca-b0ff-815eba456ebd-kube-api-access-dm2ft\") pod \"4e99e039-6910-4bca-b0ff-815eba456ebd\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.733359 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-scripts\") pod \"4e99e039-6910-4bca-b0ff-815eba456ebd\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.733457 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-config-data\") pod \"4e99e039-6910-4bca-b0ff-815eba456ebd\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.733904 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-combined-ca-bundle\") pod \"4e99e039-6910-4bca-b0ff-815eba456ebd\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.733930 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-secret-key\") pod \"4e99e039-6910-4bca-b0ff-815eba456ebd\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.733973 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e99e039-6910-4bca-b0ff-815eba456ebd-logs\") pod \"4e99e039-6910-4bca-b0ff-815eba456ebd\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.734022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-tls-certs\") pod \"4e99e039-6910-4bca-b0ff-815eba456ebd\" (UID: \"4e99e039-6910-4bca-b0ff-815eba456ebd\") " Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.735325 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e99e039-6910-4bca-b0ff-815eba456ebd-logs" (OuterVolumeSpecName: "logs") pod "4e99e039-6910-4bca-b0ff-815eba456ebd" (UID: "4e99e039-6910-4bca-b0ff-815eba456ebd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.749877 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e99e039-6910-4bca-b0ff-815eba456ebd-kube-api-access-dm2ft" (OuterVolumeSpecName: "kube-api-access-dm2ft") pod "4e99e039-6910-4bca-b0ff-815eba456ebd" (UID: "4e99e039-6910-4bca-b0ff-815eba456ebd"). InnerVolumeSpecName "kube-api-access-dm2ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.750662 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4e99e039-6910-4bca-b0ff-815eba456ebd" (UID: "4e99e039-6910-4bca-b0ff-815eba456ebd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.791963 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e99e039-6910-4bca-b0ff-815eba456ebd" (UID: "4e99e039-6910-4bca-b0ff-815eba456ebd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.800931 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-config-data" (OuterVolumeSpecName: "config-data") pod "4e99e039-6910-4bca-b0ff-815eba456ebd" (UID: "4e99e039-6910-4bca-b0ff-815eba456ebd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.813276 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-scripts" (OuterVolumeSpecName: "scripts") pod "4e99e039-6910-4bca-b0ff-815eba456ebd" (UID: "4e99e039-6910-4bca-b0ff-815eba456ebd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.829910 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6f3a5305-f80b-4798-9444-90152292e039","Type":"ContainerStarted","Data":"9e390c3243fdbc50edb03d4201efd8e7bc705d890f0f3c40703ff960ab333154"} Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.831691 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "4e99e039-6910-4bca-b0ff-815eba456ebd" (UID: "4e99e039-6910-4bca-b0ff-815eba456ebd"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.838946 4945 generic.go:334] "Generic (PLEG): container finished" podID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerID="39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87" exitCode=137 Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.839027 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58798f7484-jkqnb" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.839043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58798f7484-jkqnb" event={"ID":"4e99e039-6910-4bca-b0ff-815eba456ebd","Type":"ContainerDied","Data":"39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87"} Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.839073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58798f7484-jkqnb" event={"ID":"4e99e039-6910-4bca-b0ff-815eba456ebd","Type":"ContainerDied","Data":"1ed999df9b090cc5d890c6dce8857588abd68afdb071c73d457198c564b51b0c"} Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.839091 4945 scope.go:117] "RemoveContainer" containerID="1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.845454 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.845487 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm2ft\" (UniqueName: \"kubernetes.io/projected/4e99e039-6910-4bca-b0ff-815eba456ebd-kube-api-access-dm2ft\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.845500 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.845508 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e99e039-6910-4bca-b0ff-815eba456ebd-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.845516 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.845524 4945 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e99e039-6910-4bca-b0ff-815eba456ebd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.845532 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e99e039-6910-4bca-b0ff-815eba456ebd-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:31 crc kubenswrapper[4945]: I1008 15:35:31.847361 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bd5462d-d260-44d3-ab01-765e4fe34744","Type":"ContainerStarted","Data":"fd38417c6ef46c31a8858a42685754d130e3b94bb7cf5cd87464d2dadef67b98"} Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.044001 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62c8d55c-c623-4d14-8117-91da2634ad6b" path="/var/lib/kubelet/pods/62c8d55c-c623-4d14-8117-91da2634ad6b/volumes" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.044776 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b00bb163-6180-4acb-a485-095a289db80a" path="/var/lib/kubelet/pods/b00bb163-6180-4acb-a485-095a289db80a/volumes" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.045478 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58798f7484-jkqnb"] Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.045505 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-58798f7484-jkqnb"] Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.107234 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.185488 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.193253 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.194133 4945 scope.go:117] "RemoveContainer" containerID="a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.197895 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.235245 4945 scope.go:117] "RemoveContainer" containerID="39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.253215 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.274650 4945 scope.go:117] "RemoveContainer" containerID="1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b" Oct 08 15:35:32 crc kubenswrapper[4945]: E1008 15:35:32.275230 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b\": container with ID starting with 1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b not found: ID does not exist" containerID="1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.275255 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b"} err="failed to get container status \"1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b\": rpc error: code = NotFound desc = could not find container \"1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b\": container with ID starting with 1ba2fead1fbf381edc38bac03c18da630dc428e828e645f8886457ad9355642b not found: ID does not exist" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.275293 4945 scope.go:117] "RemoveContainer" containerID="39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87" Oct 08 15:35:32 crc kubenswrapper[4945]: E1008 15:35:32.293635 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87\": container with ID starting with 39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87 not found: ID does not exist" containerID="39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.293693 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87"} err="failed to get container status \"39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87\": rpc error: code = NotFound desc = could not find container \"39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87\": container with ID starting with 39c1e727a58143dab1ebca190f6c222e82dc87f94894c07f4f46f5f5f2ff1d87 not found: ID does not exist" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.449683 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.894670 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bd5462d-d260-44d3-ab01-765e4fe34744","Type":"ContainerStarted","Data":"46ced4d7ebc47c52e1fb94efcdfb5b3c3afe52a9f4d72c3339cea686b3bc26e0"} Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.895280 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-log" containerID="cri-o://fd38417c6ef46c31a8858a42685754d130e3b94bb7cf5cd87464d2dadef67b98" gracePeriod=30 Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.895587 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-httpd" containerID="cri-o://46ced4d7ebc47c52e1fb94efcdfb5b3c3afe52a9f4d72c3339cea686b3bc26e0" gracePeriod=30 Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.913090 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerStarted","Data":"6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20"} Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.933806 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.933788532 podStartE2EDuration="5.933788532s" podCreationTimestamp="2025-10-08 15:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:32.929357201 +0000 UTC m=+1342.283272092" watchObservedRunningTime="2025-10-08 15:35:32.933788532 +0000 UTC m=+1342.287703433" Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.945141 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6f3a5305-f80b-4798-9444-90152292e039","Type":"ContainerStarted","Data":"305d0cda5ee94e95efdbedd6f7f48b8d533c3afbcad085b02e420045d93135b3"} Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.945314 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-log" containerID="cri-o://9e390c3243fdbc50edb03d4201efd8e7bc705d890f0f3c40703ff960ab333154" gracePeriod=30 Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.945559 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-httpd" containerID="cri-o://305d0cda5ee94e95efdbedd6f7f48b8d533c3afbcad085b02e420045d93135b3" gracePeriod=30 Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.967407 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"815a3456-3316-4348-a417-f274fda65c41","Type":"ContainerStarted","Data":"e792e7fa65947bf3501f59201b0b08f67f3e2afc2587ca7fe07e720b0f65ca60"} Oct 08 15:35:32 crc kubenswrapper[4945]: I1008 15:35:32.992727 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerStarted","Data":"ce4a9b4d452d56728b2581c57a81aa89c19719b8d05fdc93a75e03a19afc5f43"} Oct 08 15:35:33 crc kubenswrapper[4945]: I1008 15:35:33.001872 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.001852962 podStartE2EDuration="6.001852962s" podCreationTimestamp="2025-10-08 15:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:32.994918518 +0000 UTC m=+1342.348833419" watchObservedRunningTime="2025-10-08 15:35:33.001852962 +0000 UTC m=+1342.355767863" Oct 08 15:35:33 crc kubenswrapper[4945]: E1008 15:35:33.169839 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f3a5305_f80b_4798_9444_90152292e039.slice/crio-conmon-305d0cda5ee94e95efdbedd6f7f48b8d533c3afbcad085b02e420045d93135b3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f3a5305_f80b_4798_9444_90152292e039.slice/crio-9e390c3243fdbc50edb03d4201efd8e7bc705d890f0f3c40703ff960ab333154.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bd5462d_d260_44d3_ab01_765e4fe34744.slice/crio-conmon-fd38417c6ef46c31a8858a42685754d130e3b94bb7cf5cd87464d2dadef67b98.scope\": RecentStats: unable to find data in memory cache]" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.020888 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerStarted","Data":"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.021360 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerStarted","Data":"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.036160 4945 generic.go:334] "Generic (PLEG): container finished" podID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerID="46ced4d7ebc47c52e1fb94efcdfb5b3c3afe52a9f4d72c3339cea686b3bc26e0" exitCode=0 Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.037146 4945 generic.go:334] "Generic (PLEG): container finished" podID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerID="fd38417c6ef46c31a8858a42685754d130e3b94bb7cf5cd87464d2dadef67b98" exitCode=143 Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.044975 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" path="/var/lib/kubelet/pods/4e99e039-6910-4bca-b0ff-815eba456ebd/volumes" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.045778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bd5462d-d260-44d3-ab01-765e4fe34744","Type":"ContainerDied","Data":"46ced4d7ebc47c52e1fb94efcdfb5b3c3afe52a9f4d72c3339cea686b3bc26e0"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.045997 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bd5462d-d260-44d3-ab01-765e4fe34744","Type":"ContainerDied","Data":"fd38417c6ef46c31a8858a42685754d130e3b94bb7cf5cd87464d2dadef67b98"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.054264 4945 generic.go:334] "Generic (PLEG): container finished" podID="6f3a5305-f80b-4798-9444-90152292e039" containerID="305d0cda5ee94e95efdbedd6f7f48b8d533c3afbcad085b02e420045d93135b3" exitCode=143 Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.054410 4945 generic.go:334] "Generic (PLEG): container finished" podID="6f3a5305-f80b-4798-9444-90152292e039" containerID="9e390c3243fdbc50edb03d4201efd8e7bc705d890f0f3c40703ff960ab333154" exitCode=143 Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.054509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6f3a5305-f80b-4798-9444-90152292e039","Type":"ContainerDied","Data":"305d0cda5ee94e95efdbedd6f7f48b8d533c3afbcad085b02e420045d93135b3"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.054612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6f3a5305-f80b-4798-9444-90152292e039","Type":"ContainerDied","Data":"9e390c3243fdbc50edb03d4201efd8e7bc705d890f0f3c40703ff960ab333154"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.054673 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6f3a5305-f80b-4798-9444-90152292e039","Type":"ContainerDied","Data":"ec68f09d4c01517e4795689d35bc937538db86efdf08cedf06a26d50ce9d1c17"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.054729 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec68f09d4c01517e4795689d35bc937538db86efdf08cedf06a26d50ce9d1c17" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.069771 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"815a3456-3316-4348-a417-f274fda65c41","Type":"ContainerStarted","Data":"8963ae0b277d95ca236d5ddf2c9ca2a044c292485e5d3498fe8d1a632e28803b"} Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.095775 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.112329 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.229617 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.229727 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-httpd-run\") pod \"2bd5462d-d260-44d3-ab01-765e4fe34744\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.229765 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-scripts\") pod \"2bd5462d-d260-44d3-ab01-765e4fe34744\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.235641 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2bd5462d-d260-44d3-ab01-765e4fe34744" (UID: "2bd5462d-d260-44d3-ab01-765e4fe34744"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236136 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-config-data\") pod \"2bd5462d-d260-44d3-ab01-765e4fe34744\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236205 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-logs\") pod \"2bd5462d-d260-44d3-ab01-765e4fe34744\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236236 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-scripts\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236303 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-combined-ca-bundle\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236334 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-combined-ca-bundle\") pod \"2bd5462d-d260-44d3-ab01-765e4fe34744\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xbjr\" (UniqueName: \"kubernetes.io/projected/2bd5462d-d260-44d3-ab01-765e4fe34744-kube-api-access-7xbjr\") pod \"2bd5462d-d260-44d3-ab01-765e4fe34744\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236379 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236465 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-httpd-run\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.236479 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-logs\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.237177 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"2bd5462d-d260-44d3-ab01-765e4fe34744\" (UID: \"2bd5462d-d260-44d3-ab01-765e4fe34744\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.237255 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w9hm\" (UniqueName: \"kubernetes.io/projected/6f3a5305-f80b-4798-9444-90152292e039-kube-api-access-8w9hm\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.237961 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.239261 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-scripts" (OuterVolumeSpecName: "scripts") pod "2bd5462d-d260-44d3-ab01-765e4fe34744" (UID: "2bd5462d-d260-44d3-ab01-765e4fe34744"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.241777 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "2bd5462d-d260-44d3-ab01-765e4fe34744" (UID: "2bd5462d-d260-44d3-ab01-765e4fe34744"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.248580 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-logs" (OuterVolumeSpecName: "logs") pod "2bd5462d-d260-44d3-ab01-765e4fe34744" (UID: "2bd5462d-d260-44d3-ab01-765e4fe34744"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.250256 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-logs" (OuterVolumeSpecName: "logs") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.250331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.255437 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f3a5305-f80b-4798-9444-90152292e039-kube-api-access-8w9hm" (OuterVolumeSpecName: "kube-api-access-8w9hm") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "kube-api-access-8w9hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.259255 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd5462d-d260-44d3-ab01-765e4fe34744-kube-api-access-7xbjr" (OuterVolumeSpecName: "kube-api-access-7xbjr") pod "2bd5462d-d260-44d3-ab01-765e4fe34744" (UID: "2bd5462d-d260-44d3-ab01-765e4fe34744"). InnerVolumeSpecName "kube-api-access-7xbjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.259290 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-scripts" (OuterVolumeSpecName: "scripts") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.259290 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.280691 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.282552 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bd5462d-d260-44d3-ab01-765e4fe34744" (UID: "2bd5462d-d260-44d3-ab01-765e4fe34744"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data" (OuterVolumeSpecName: "config-data") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346456 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data\") pod \"6f3a5305-f80b-4798-9444-90152292e039\" (UID: \"6f3a5305-f80b-4798-9444-90152292e039\") " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346893 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346913 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f3a5305-f80b-4798-9444-90152292e039-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346933 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346942 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w9hm\" (UniqueName: \"kubernetes.io/projected/6f3a5305-f80b-4798-9444-90152292e039-kube-api-access-8w9hm\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346953 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346961 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd5462d-d260-44d3-ab01-765e4fe34744-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346969 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346977 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346986 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.346994 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xbjr\" (UniqueName: \"kubernetes.io/projected/2bd5462d-d260-44d3-ab01-765e4fe34744-kube-api-access-7xbjr\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.347007 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 08 15:35:34 crc kubenswrapper[4945]: W1008 15:35:34.347223 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/6f3a5305-f80b-4798-9444-90152292e039/volumes/kubernetes.io~secret/config-data Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.347249 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data" (OuterVolumeSpecName: "config-data") pod "6f3a5305-f80b-4798-9444-90152292e039" (UID: "6f3a5305-f80b-4798-9444-90152292e039"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.362496 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.374972 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-config-data" (OuterVolumeSpecName: "config-data") pod "2bd5462d-d260-44d3-ab01-765e4fe34744" (UID: "2bd5462d-d260-44d3-ab01-765e4fe34744"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.382188 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.393640 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.452063 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f3a5305-f80b-4798-9444-90152292e039-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.452096 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd5462d-d260-44d3-ab01-765e4fe34744-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.452133 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:34 crc kubenswrapper[4945]: I1008 15:35:34.452146 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.080132 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerStarted","Data":"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5"} Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.083060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bd5462d-d260-44d3-ab01-765e4fe34744","Type":"ContainerDied","Data":"f63d4e19d792cb0374f3e01b042a3eebd4fb63845c6d6d6e33fb043c3748eeca"} Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.083094 4945 scope.go:117] "RemoveContainer" containerID="46ced4d7ebc47c52e1fb94efcdfb5b3c3afe52a9f4d72c3339cea686b3bc26e0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.083116 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.085172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"815a3456-3316-4348-a417-f274fda65c41","Type":"ContainerStarted","Data":"44e18acba8eb28fb8ae44ee081d7b6856cf6d1b31422112c826d9e5df01fd9a5"} Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.085194 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.110744 4945 scope.go:117] "RemoveContainer" containerID="fd38417c6ef46c31a8858a42685754d130e3b94bb7cf5cd87464d2dadef67b98" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.119891 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.130052 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.138426 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.145141 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160165 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: E1008 15:35:35.160574 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160591 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" Oct 08 15:35:35 crc kubenswrapper[4945]: E1008 15:35:35.160616 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-log" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160622 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-log" Oct 08 15:35:35 crc kubenswrapper[4945]: E1008 15:35:35.160631 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-log" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160637 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-log" Oct 08 15:35:35 crc kubenswrapper[4945]: E1008 15:35:35.160651 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-httpd" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160657 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-httpd" Oct 08 15:35:35 crc kubenswrapper[4945]: E1008 15:35:35.160674 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon-log" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160680 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon-log" Oct 08 15:35:35 crc kubenswrapper[4945]: E1008 15:35:35.160687 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-httpd" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160692 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-httpd" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160856 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-httpd" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160876 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f3a5305-f80b-4798-9444-90152292e039" containerName="glance-log" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160884 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160894 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-httpd" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160906 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e99e039-6910-4bca-b0ff-815eba456ebd" containerName="horizon-log" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.160916 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" containerName="glance-log" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.161928 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.164291 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.164452 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g97m6" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.164603 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.164816 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.169036 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.175047 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.178346 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.178513 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.234064 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.253524 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.265903 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.265955 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.265983 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-scripts\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266023 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-config-data\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266054 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266069 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhcpg\" (UniqueName: \"kubernetes.io/projected/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-kube-api-access-jhcpg\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266104 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266175 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs22t\" (UniqueName: \"kubernetes.io/projected/be98f960-821f-4340-9d7c-6cb26b2b330f-kube-api-access-cs22t\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266199 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266217 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266235 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-logs\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266257 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266277 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266310 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-logs\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266327 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.266359 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-logs\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368496 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368523 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368566 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-scripts\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368604 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-config-data\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368634 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368657 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhcpg\" (UniqueName: \"kubernetes.io/projected/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-kube-api-access-jhcpg\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368706 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368730 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs22t\" (UniqueName: \"kubernetes.io/projected/be98f960-821f-4340-9d7c-6cb26b2b330f-kube-api-access-cs22t\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368816 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-logs\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368852 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.368886 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.369390 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.370286 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.370801 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-logs\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.371124 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-logs\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.371660 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.370532 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.389179 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.392451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.392894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.393974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.400530 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.404702 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-scripts\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.408550 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.414280 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-config-data\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.426671 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs22t\" (UniqueName: \"kubernetes.io/projected/be98f960-821f-4340-9d7c-6cb26b2b330f-kube-api-access-cs22t\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.431472 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhcpg\" (UniqueName: \"kubernetes.io/projected/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-kube-api-access-jhcpg\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.443623 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.473282 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.503618 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.511921 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:35 crc kubenswrapper[4945]: I1008 15:35:35.788831 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:36 crc kubenswrapper[4945]: I1008 15:35:36.110369 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd5462d-d260-44d3-ab01-765e4fe34744" path="/var/lib/kubelet/pods/2bd5462d-d260-44d3-ab01-765e4fe34744/volumes" Oct 08 15:35:36 crc kubenswrapper[4945]: I1008 15:35:36.116496 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f3a5305-f80b-4798-9444-90152292e039" path="/var/lib/kubelet/pods/6f3a5305-f80b-4798-9444-90152292e039/volumes" Oct 08 15:35:36 crc kubenswrapper[4945]: I1008 15:35:36.158773 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.15875563 podStartE2EDuration="5.15875563s" podCreationTimestamp="2025-10-08 15:35:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:36.152963885 +0000 UTC m=+1345.506878786" watchObservedRunningTime="2025-10-08 15:35:36.15875563 +0000 UTC m=+1345.512670531" Oct 08 15:35:36 crc kubenswrapper[4945]: I1008 15:35:36.595196 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 15:35:37 crc kubenswrapper[4945]: I1008 15:35:37.540051 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:37 crc kubenswrapper[4945]: I1008 15:35:37.542898 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" Oct 08 15:35:37 crc kubenswrapper[4945]: I1008 15:35:37.695478 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:37 crc kubenswrapper[4945]: I1008 15:35:37.695527 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66d5fb75f8-44p9s" Oct 08 15:35:38 crc kubenswrapper[4945]: I1008 15:35:38.070365 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:35:38 crc kubenswrapper[4945]: I1008 15:35:38.139892 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f57676dc-brk7t"] Oct 08 15:35:38 crc kubenswrapper[4945]: I1008 15:35:38.140151 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" podUID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerName="dnsmasq-dns" containerID="cri-o://0d7f328d4908a76cb92f7f05e073bbad433dd5de97057a23c56a54975d3a98ef" gracePeriod=10 Oct 08 15:35:38 crc kubenswrapper[4945]: I1008 15:35:38.174954 4945 generic.go:334] "Generic (PLEG): container finished" podID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" exitCode=1 Oct 08 15:35:38 crc kubenswrapper[4945]: I1008 15:35:38.175593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerDied","Data":"6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20"} Oct 08 15:35:38 crc kubenswrapper[4945]: I1008 15:35:38.175656 4945 scope.go:117] "RemoveContainer" containerID="a2f7febdd7db33d84303893be5203ac36c8c53828ca7b4b48dce2d55b24251b6" Oct 08 15:35:38 crc kubenswrapper[4945]: I1008 15:35:38.177699 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:35:38 crc kubenswrapper[4945]: E1008 15:35:38.178065 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:35:39 crc kubenswrapper[4945]: I1008 15:35:39.184789 4945 generic.go:334] "Generic (PLEG): container finished" podID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerID="0d7f328d4908a76cb92f7f05e073bbad433dd5de97057a23c56a54975d3a98ef" exitCode=0 Oct 08 15:35:39 crc kubenswrapper[4945]: I1008 15:35:39.184838 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" event={"ID":"4cdde67a-2aad-4ef4-ae1b-db394a522f32","Type":"ContainerDied","Data":"0d7f328d4908a76cb92f7f05e073bbad433dd5de97057a23c56a54975d3a98ef"} Oct 08 15:35:41 crc kubenswrapper[4945]: I1008 15:35:41.764491 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.193226 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.193286 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.194056 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:35:42 crc kubenswrapper[4945]: E1008 15:35:42.194359 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.717989 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.820203 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-svc\") pod \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.820290 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-nb\") pod \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.820319 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-config\") pod \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.820383 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-sb\") pod \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.820537 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-swift-storage-0\") pod \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.820563 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhj94\" (UniqueName: \"kubernetes.io/projected/4cdde67a-2aad-4ef4-ae1b-db394a522f32-kube-api-access-hhj94\") pod \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\" (UID: \"4cdde67a-2aad-4ef4-ae1b-db394a522f32\") " Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.826747 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cdde67a-2aad-4ef4-ae1b-db394a522f32-kube-api-access-hhj94" (OuterVolumeSpecName: "kube-api-access-hhj94") pod "4cdde67a-2aad-4ef4-ae1b-db394a522f32" (UID: "4cdde67a-2aad-4ef4-ae1b-db394a522f32"). InnerVolumeSpecName "kube-api-access-hhj94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.875716 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4cdde67a-2aad-4ef4-ae1b-db394a522f32" (UID: "4cdde67a-2aad-4ef4-ae1b-db394a522f32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.888844 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4cdde67a-2aad-4ef4-ae1b-db394a522f32" (UID: "4cdde67a-2aad-4ef4-ae1b-db394a522f32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.901366 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-config" (OuterVolumeSpecName: "config") pod "4cdde67a-2aad-4ef4-ae1b-db394a522f32" (UID: "4cdde67a-2aad-4ef4-ae1b-db394a522f32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.901662 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4cdde67a-2aad-4ef4-ae1b-db394a522f32" (UID: "4cdde67a-2aad-4ef4-ae1b-db394a522f32"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.908313 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4cdde67a-2aad-4ef4-ae1b-db394a522f32" (UID: "4cdde67a-2aad-4ef4-ae1b-db394a522f32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.923014 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.923042 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.923053 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.923062 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.923072 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cdde67a-2aad-4ef4-ae1b-db394a522f32-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:42 crc kubenswrapper[4945]: I1008 15:35:42.923080 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhj94\" (UniqueName: \"kubernetes.io/projected/4cdde67a-2aad-4ef4-ae1b-db394a522f32-kube-api-access-hhj94\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.068367 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:43 crc kubenswrapper[4945]: W1008 15:35:43.069987 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bf19d0f_a799_43f5_b39c_c9d98fdf0e41.slice/crio-251a36aa8e95d984750d1c7613aef7dc9d260fd99c89a358205b6e7d5b254081 WatchSource:0}: Error finding container 251a36aa8e95d984750d1c7613aef7dc9d260fd99c89a358205b6e7d5b254081: Status 404 returned error can't find the container with id 251a36aa8e95d984750d1c7613aef7dc9d260fd99c89a358205b6e7d5b254081 Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.130407 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5z5q5"] Oct 08 15:35:43 crc kubenswrapper[4945]: E1008 15:35:43.130886 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerName="init" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.130904 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerName="init" Oct 08 15:35:43 crc kubenswrapper[4945]: E1008 15:35:43.130923 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerName="dnsmasq-dns" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.130929 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerName="dnsmasq-dns" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.131140 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" containerName="dnsmasq-dns" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.131784 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z5q5" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.142396 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5z5q5"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.229034 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-z7bqr"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.229837 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2v6s\" (UniqueName: \"kubernetes.io/projected/6a4bdf24-ed65-4aa7-acd9-922bd48a1b88-kube-api-access-r2v6s\") pod \"nova-api-db-create-5z5q5\" (UID: \"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88\") " pod="openstack/nova-api-db-create-5z5q5" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.231768 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z7bqr" Oct 08 15:35:43 crc kubenswrapper[4945]: W1008 15:35:43.253854 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe98f960_821f_4340_9d7c_6cb26b2b330f.slice/crio-282b8867efb95f43a1e609a575bac3ee062b1a10fb2daab1a5915996ef1f7b4d WatchSource:0}: Error finding container 282b8867efb95f43a1e609a575bac3ee062b1a10fb2daab1a5915996ef1f7b4d: Status 404 returned error can't find the container with id 282b8867efb95f43a1e609a575bac3ee062b1a10fb2daab1a5915996ef1f7b4d Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.260719 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" event={"ID":"4cdde67a-2aad-4ef4-ae1b-db394a522f32","Type":"ContainerDied","Data":"b56415ed8bc5ff59a2bc104085cfeb659fd09b8c11a8d36d28d9190e0d1011f9"} Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.260773 4945 scope.go:117] "RemoveContainer" containerID="0d7f328d4908a76cb92f7f05e073bbad433dd5de97057a23c56a54975d3a98ef" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.260925 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f57676dc-brk7t" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.267846 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"223e741f-c7cc-465f-a4bc-804bfff7a6ef","Type":"ContainerStarted","Data":"cb55fbfafec01fd19b625ea0f80ba17db24d4c5f751d695615dd8ac1f41f1484"} Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.274966 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z7bqr"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.287461 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41","Type":"ContainerStarted","Data":"251a36aa8e95d984750d1c7613aef7dc9d260fd99c89a358205b6e7d5b254081"} Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.293931 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.299908 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerStarted","Data":"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726"} Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.300083 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-central-agent" containerID="cri-o://31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9" gracePeriod=30 Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.300179 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.300214 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="proxy-httpd" containerID="cri-o://8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726" gracePeriod=30 Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.300250 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="sg-core" containerID="cri-o://0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5" gracePeriod=30 Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.300281 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-notification-agent" containerID="cri-o://31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807" gracePeriod=30 Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.321436 4945 scope.go:117] "RemoveContainer" containerID="76b9f3d811279210459af5f4fe10da8845675fa8f383ef3db696a2101a0b5484" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.327994 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.918821826 podStartE2EDuration="26.327975609s" podCreationTimestamp="2025-10-08 15:35:17 +0000 UTC" firstStartedPulling="2025-10-08 15:35:19.074254608 +0000 UTC m=+1328.428169509" lastFinishedPulling="2025-10-08 15:35:42.483408391 +0000 UTC m=+1351.837323292" observedRunningTime="2025-10-08 15:35:43.286931277 +0000 UTC m=+1352.640846178" watchObservedRunningTime="2025-10-08 15:35:43.327975609 +0000 UTC m=+1352.681890510" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.331740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbzqt\" (UniqueName: \"kubernetes.io/projected/67017489-fd71-4e43-99c6-7918bc9f41f5-kube-api-access-dbzqt\") pod \"nova-cell0-db-create-z7bqr\" (UID: \"67017489-fd71-4e43-99c6-7918bc9f41f5\") " pod="openstack/nova-cell0-db-create-z7bqr" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.331869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2v6s\" (UniqueName: \"kubernetes.io/projected/6a4bdf24-ed65-4aa7-acd9-922bd48a1b88-kube-api-access-r2v6s\") pod \"nova-api-db-create-5z5q5\" (UID: \"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88\") " pod="openstack/nova-api-db-create-5z5q5" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.357145 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2v6s\" (UniqueName: \"kubernetes.io/projected/6a4bdf24-ed65-4aa7-acd9-922bd48a1b88-kube-api-access-r2v6s\") pod \"nova-api-db-create-5z5q5\" (UID: \"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88\") " pod="openstack/nova-api-db-create-5z5q5" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.364799 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f57676dc-brk7t"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.375894 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f57676dc-brk7t"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.380259 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.342143901 podStartE2EDuration="12.380240472s" podCreationTimestamp="2025-10-08 15:35:31 +0000 UTC" firstStartedPulling="2025-10-08 15:35:32.477699984 +0000 UTC m=+1341.831614885" lastFinishedPulling="2025-10-08 15:35:42.515796555 +0000 UTC m=+1351.869711456" observedRunningTime="2025-10-08 15:35:43.343743505 +0000 UTC m=+1352.697658406" watchObservedRunningTime="2025-10-08 15:35:43.380240472 +0000 UTC m=+1352.734155373" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.431715 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-df5bp"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.433829 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbzqt\" (UniqueName: \"kubernetes.io/projected/67017489-fd71-4e43-99c6-7918bc9f41f5-kube-api-access-dbzqt\") pod \"nova-cell0-db-create-z7bqr\" (UID: \"67017489-fd71-4e43-99c6-7918bc9f41f5\") " pod="openstack/nova-cell0-db-create-z7bqr" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.434283 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df5bp" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.459814 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-df5bp"] Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.460667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbzqt\" (UniqueName: \"kubernetes.io/projected/67017489-fd71-4e43-99c6-7918bc9f41f5-kube-api-access-dbzqt\") pod \"nova-cell0-db-create-z7bqr\" (UID: \"67017489-fd71-4e43-99c6-7918bc9f41f5\") " pod="openstack/nova-cell0-db-create-z7bqr" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.461172 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z5q5" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.535720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbwz7\" (UniqueName: \"kubernetes.io/projected/f69da577-0161-4756-94be-12ee983945bd-kube-api-access-vbwz7\") pod \"nova-cell1-db-create-df5bp\" (UID: \"f69da577-0161-4756-94be-12ee983945bd\") " pod="openstack/nova-cell1-db-create-df5bp" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.550329 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z7bqr" Oct 08 15:35:43 crc kubenswrapper[4945]: E1008 15:35:43.640479 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode10c8660_ea2f_483c_afbe_7f7e6287968c.slice/crio-0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cdde67a_2aad_4ef4_ae1b_db394a522f32.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cdde67a_2aad_4ef4_ae1b_db394a522f32.slice/crio-b56415ed8bc5ff59a2bc104085cfeb659fd09b8c11a8d36d28d9190e0d1011f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode10c8660_ea2f_483c_afbe_7f7e6287968c.slice/crio-conmon-0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode10c8660_ea2f_483c_afbe_7f7e6287968c.slice/crio-8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726.scope\": RecentStats: unable to find data in memory cache]" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.647234 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbwz7\" (UniqueName: \"kubernetes.io/projected/f69da577-0161-4756-94be-12ee983945bd-kube-api-access-vbwz7\") pod \"nova-cell1-db-create-df5bp\" (UID: \"f69da577-0161-4756-94be-12ee983945bd\") " pod="openstack/nova-cell1-db-create-df5bp" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.703241 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbwz7\" (UniqueName: \"kubernetes.io/projected/f69da577-0161-4756-94be-12ee983945bd-kube-api-access-vbwz7\") pod \"nova-cell1-db-create-df5bp\" (UID: \"f69da577-0161-4756-94be-12ee983945bd\") " pod="openstack/nova-cell1-db-create-df5bp" Oct 08 15:35:43 crc kubenswrapper[4945]: I1008 15:35:43.886335 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df5bp" Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.041770 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cdde67a-2aad-4ef4-ae1b-db394a522f32" path="/var/lib/kubelet/pods/4cdde67a-2aad-4ef4-ae1b-db394a522f32/volumes" Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.179302 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z7bqr"] Oct 08 15:35:44 crc kubenswrapper[4945]: W1008 15:35:44.187986 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67017489_fd71_4e43_99c6_7918bc9f41f5.slice/crio-f35b2148706eddd5194cd6b276978f42649b76ef06976b74a80cf8e0de6c7ea9 WatchSource:0}: Error finding container f35b2148706eddd5194cd6b276978f42649b76ef06976b74a80cf8e0de6c7ea9: Status 404 returned error can't find the container with id f35b2148706eddd5194cd6b276978f42649b76ef06976b74a80cf8e0de6c7ea9 Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.292883 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5z5q5"] Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.317166 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41","Type":"ContainerStarted","Data":"e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8"} Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.321328 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z7bqr" event={"ID":"67017489-fd71-4e43-99c6-7918bc9f41f5","Type":"ContainerStarted","Data":"f35b2148706eddd5194cd6b276978f42649b76ef06976b74a80cf8e0de6c7ea9"} Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.327660 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be98f960-821f-4340-9d7c-6cb26b2b330f","Type":"ContainerStarted","Data":"282b8867efb95f43a1e609a575bac3ee062b1a10fb2daab1a5915996ef1f7b4d"} Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.343281 4945 generic.go:334] "Generic (PLEG): container finished" podID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerID="8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726" exitCode=0 Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.343320 4945 generic.go:334] "Generic (PLEG): container finished" podID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerID="0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5" exitCode=2 Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.343686 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerDied","Data":"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726"} Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.343738 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerDied","Data":"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5"} Oct 08 15:35:44 crc kubenswrapper[4945]: I1008 15:35:44.527833 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-df5bp"] Oct 08 15:35:44 crc kubenswrapper[4945]: W1008 15:35:44.546825 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf69da577_0161_4756_94be_12ee983945bd.slice/crio-91e3d42624467e2f64f8614ac8a9f080d315ea947f5544be62385500251f5a77 WatchSource:0}: Error finding container 91e3d42624467e2f64f8614ac8a9f080d315ea947f5544be62385500251f5a77: Status 404 returned error can't find the container with id 91e3d42624467e2f64f8614ac8a9f080d315ea947f5544be62385500251f5a77 Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.401624 4945 generic.go:334] "Generic (PLEG): container finished" podID="f69da577-0161-4756-94be-12ee983945bd" containerID="8914d335e229a76e0b5acc4ba67784634f85d1084b9068152005b837fc496c79" exitCode=0 Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.402515 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-df5bp" event={"ID":"f69da577-0161-4756-94be-12ee983945bd","Type":"ContainerDied","Data":"8914d335e229a76e0b5acc4ba67784634f85d1084b9068152005b837fc496c79"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.402549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-df5bp" event={"ID":"f69da577-0161-4756-94be-12ee983945bd","Type":"ContainerStarted","Data":"91e3d42624467e2f64f8614ac8a9f080d315ea947f5544be62385500251f5a77"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.413345 4945 generic.go:334] "Generic (PLEG): container finished" podID="6a4bdf24-ed65-4aa7-acd9-922bd48a1b88" containerID="132a5d7c75f48d809e36d017e92f44ede7220c89ea8f80b926eff721625df2dc" exitCode=0 Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.413435 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5z5q5" event={"ID":"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88","Type":"ContainerDied","Data":"132a5d7c75f48d809e36d017e92f44ede7220c89ea8f80b926eff721625df2dc"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.413464 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5z5q5" event={"ID":"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88","Type":"ContainerStarted","Data":"82101d0ef334954794dec5caf10128501d7fd3002adf9b717e9ddb1767fa225e"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.429841 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41","Type":"ContainerStarted","Data":"02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.433367 4945 generic.go:334] "Generic (PLEG): container finished" podID="67017489-fd71-4e43-99c6-7918bc9f41f5" containerID="de4fc6dd04fd18b220a8754fb329f78ade3462fabcbbeb2d90071fe1eea1787d" exitCode=0 Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.433436 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z7bqr" event={"ID":"67017489-fd71-4e43-99c6-7918bc9f41f5","Type":"ContainerDied","Data":"de4fc6dd04fd18b220a8754fb329f78ade3462fabcbbeb2d90071fe1eea1787d"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.435671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be98f960-821f-4340-9d7c-6cb26b2b330f","Type":"ContainerStarted","Data":"a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.435704 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be98f960-821f-4340-9d7c-6cb26b2b330f","Type":"ContainerStarted","Data":"c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77"} Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.482374 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.482350111 podStartE2EDuration="10.482350111s" podCreationTimestamp="2025-10-08 15:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:45.465645211 +0000 UTC m=+1354.819560112" watchObservedRunningTime="2025-10-08 15:35:45.482350111 +0000 UTC m=+1354.836265012" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.501236 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.501210616 podStartE2EDuration="10.501210616s" podCreationTimestamp="2025-10-08 15:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:45.49303623 +0000 UTC m=+1354.846951141" watchObservedRunningTime="2025-10-08 15:35:45.501210616 +0000 UTC m=+1354.855125517" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.504949 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.504991 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.534348 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.545315 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.789573 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.789736 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.818906 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 15:35:45 crc kubenswrapper[4945]: I1008 15:35:45.846958 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.176818 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.299347 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-log-httpd\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.299388 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.299445 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-run-httpd\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.299510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-sg-core-conf-yaml\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.299592 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pdrp\" (UniqueName: \"kubernetes.io/projected/e10c8660-ea2f-483c-afbe-7f7e6287968c-kube-api-access-9pdrp\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.299616 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-combined-ca-bundle\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.299639 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-scripts\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.301328 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.301810 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.305474 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e10c8660-ea2f-483c-afbe-7f7e6287968c-kube-api-access-9pdrp" (OuterVolumeSpecName: "kube-api-access-9pdrp") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "kube-api-access-9pdrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.306832 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-scripts" (OuterVolumeSpecName: "scripts") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.331272 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.395294 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.401128 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data" (OuterVolumeSpecName: "config-data") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.401660 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data\") pod \"e10c8660-ea2f-483c-afbe-7f7e6287968c\" (UID: \"e10c8660-ea2f-483c-afbe-7f7e6287968c\") " Oct 08 15:35:46 crc kubenswrapper[4945]: W1008 15:35:46.401789 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/e10c8660-ea2f-483c-afbe-7f7e6287968c/volumes/kubernetes.io~secret/config-data Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.401808 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data" (OuterVolumeSpecName: "config-data") pod "e10c8660-ea2f-483c-afbe-7f7e6287968c" (UID: "e10c8660-ea2f-483c-afbe-7f7e6287968c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.402085 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pdrp\" (UniqueName: \"kubernetes.io/projected/e10c8660-ea2f-483c-afbe-7f7e6287968c-kube-api-access-9pdrp\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.402102 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.402153 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.402161 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.402169 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.402177 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e10c8660-ea2f-483c-afbe-7f7e6287968c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.402185 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e10c8660-ea2f-483c-afbe-7f7e6287968c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.445954 4945 generic.go:334] "Generic (PLEG): container finished" podID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerID="31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807" exitCode=0 Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.445985 4945 generic.go:334] "Generic (PLEG): container finished" podID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerID="31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9" exitCode=0 Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.446949 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.448289 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerDied","Data":"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807"} Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.448343 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerDied","Data":"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9"} Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.448354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e10c8660-ea2f-483c-afbe-7f7e6287968c","Type":"ContainerDied","Data":"ce4a9b4d452d56728b2581c57a81aa89c19719b8d05fdc93a75e03a19afc5f43"} Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.448370 4945 scope.go:117] "RemoveContainer" containerID="8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.448424 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.449852 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.449893 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.449909 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.487842 4945 scope.go:117] "RemoveContainer" containerID="0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.489218 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.501421 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.509524 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.510227 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-notification-agent" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510251 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-notification-agent" Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.510285 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-central-agent" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510296 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-central-agent" Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.510319 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="proxy-httpd" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510328 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="proxy-httpd" Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.510342 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="sg-core" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510351 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="sg-core" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510609 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-central-agent" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510638 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="proxy-httpd" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510648 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="sg-core" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.510677 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" containerName="ceilometer-notification-agent" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.513204 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.517518 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.527965 4945 scope.go:117] "RemoveContainer" containerID="31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.528412 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.528532 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.564355 4945 scope.go:117] "RemoveContainer" containerID="31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.587036 4945 scope.go:117] "RemoveContainer" containerID="8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726" Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.587742 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726\": container with ID starting with 8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726 not found: ID does not exist" containerID="8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.587779 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726"} err="failed to get container status \"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726\": rpc error: code = NotFound desc = could not find container \"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726\": container with ID starting with 8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.587806 4945 scope.go:117] "RemoveContainer" containerID="0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5" Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.588148 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5\": container with ID starting with 0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5 not found: ID does not exist" containerID="0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588182 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5"} err="failed to get container status \"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5\": rpc error: code = NotFound desc = could not find container \"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5\": container with ID starting with 0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588203 4945 scope.go:117] "RemoveContainer" containerID="31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807" Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.588431 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807\": container with ID starting with 31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807 not found: ID does not exist" containerID="31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588455 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807"} err="failed to get container status \"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807\": rpc error: code = NotFound desc = could not find container \"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807\": container with ID starting with 31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588470 4945 scope.go:117] "RemoveContainer" containerID="31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9" Oct 08 15:35:46 crc kubenswrapper[4945]: E1008 15:35:46.588653 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9\": container with ID starting with 31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9 not found: ID does not exist" containerID="31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588674 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9"} err="failed to get container status \"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9\": rpc error: code = NotFound desc = could not find container \"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9\": container with ID starting with 31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588688 4945 scope.go:117] "RemoveContainer" containerID="8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588899 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726"} err="failed to get container status \"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726\": rpc error: code = NotFound desc = could not find container \"8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726\": container with ID starting with 8ed5e240a121a499409f2442ac7d728d49f082da83ad98c3c8920fff2166d726 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.588922 4945 scope.go:117] "RemoveContainer" containerID="0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.589092 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5"} err="failed to get container status \"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5\": rpc error: code = NotFound desc = could not find container \"0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5\": container with ID starting with 0e9d28e8cd3380d5d2ade07570e3964e5b4f1c79defc1c51ffef26dc032b45c5 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.589159 4945 scope.go:117] "RemoveContainer" containerID="31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.589368 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807"} err="failed to get container status \"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807\": rpc error: code = NotFound desc = could not find container \"31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807\": container with ID starting with 31649569af78b00ceefce6fb87cc5d1bfbf66b8178f8756960382baac871c807 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.589388 4945 scope.go:117] "RemoveContainer" containerID="31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.589558 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9"} err="failed to get container status \"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9\": rpc error: code = NotFound desc = could not find container \"31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9\": container with ID starting with 31f026fce9cb6ffcee94e8e4f98200b65af3ac6a16f6440aa32ddd1e6da2d0e9 not found: ID does not exist" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.605729 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7lxn\" (UniqueName: \"kubernetes.io/projected/405d1d9b-075c-445a-9463-0d79a97338a0-kube-api-access-p7lxn\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.605784 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-log-httpd\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.605824 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-run-httpd\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.605884 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-scripts\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.605920 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.605960 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-config-data\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.606072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.707653 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-scripts\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.707921 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.707987 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-config-data\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.708231 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.708414 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7lxn\" (UniqueName: \"kubernetes.io/projected/405d1d9b-075c-445a-9463-0d79a97338a0-kube-api-access-p7lxn\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.708505 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-log-httpd\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.708578 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-run-httpd\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.711162 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-log-httpd\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.711203 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-run-httpd\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.712336 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-config-data\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.712894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-scripts\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.713714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.716430 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.735669 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7lxn\" (UniqueName: \"kubernetes.io/projected/405d1d9b-075c-445a-9463-0d79a97338a0-kube-api-access-p7lxn\") pod \"ceilometer-0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.848567 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.944615 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z5q5" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.962468 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z7bqr" Oct 08 15:35:46 crc kubenswrapper[4945]: I1008 15:35:46.980837 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df5bp" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.012956 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2v6s\" (UniqueName: \"kubernetes.io/projected/6a4bdf24-ed65-4aa7-acd9-922bd48a1b88-kube-api-access-r2v6s\") pod \"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88\" (UID: \"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88\") " Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.020720 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a4bdf24-ed65-4aa7-acd9-922bd48a1b88-kube-api-access-r2v6s" (OuterVolumeSpecName: "kube-api-access-r2v6s") pod "6a4bdf24-ed65-4aa7-acd9-922bd48a1b88" (UID: "6a4bdf24-ed65-4aa7-acd9-922bd48a1b88"). InnerVolumeSpecName "kube-api-access-r2v6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.114590 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbzqt\" (UniqueName: \"kubernetes.io/projected/67017489-fd71-4e43-99c6-7918bc9f41f5-kube-api-access-dbzqt\") pod \"67017489-fd71-4e43-99c6-7918bc9f41f5\" (UID: \"67017489-fd71-4e43-99c6-7918bc9f41f5\") " Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.114649 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbwz7\" (UniqueName: \"kubernetes.io/projected/f69da577-0161-4756-94be-12ee983945bd-kube-api-access-vbwz7\") pod \"f69da577-0161-4756-94be-12ee983945bd\" (UID: \"f69da577-0161-4756-94be-12ee983945bd\") " Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.115587 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2v6s\" (UniqueName: \"kubernetes.io/projected/6a4bdf24-ed65-4aa7-acd9-922bd48a1b88-kube-api-access-r2v6s\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.119390 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69da577-0161-4756-94be-12ee983945bd-kube-api-access-vbwz7" (OuterVolumeSpecName: "kube-api-access-vbwz7") pod "f69da577-0161-4756-94be-12ee983945bd" (UID: "f69da577-0161-4756-94be-12ee983945bd"). InnerVolumeSpecName "kube-api-access-vbwz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.119453 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67017489-fd71-4e43-99c6-7918bc9f41f5-kube-api-access-dbzqt" (OuterVolumeSpecName: "kube-api-access-dbzqt") pod "67017489-fd71-4e43-99c6-7918bc9f41f5" (UID: "67017489-fd71-4e43-99c6-7918bc9f41f5"). InnerVolumeSpecName "kube-api-access-dbzqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.216956 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbzqt\" (UniqueName: \"kubernetes.io/projected/67017489-fd71-4e43-99c6-7918bc9f41f5-kube-api-access-dbzqt\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.216990 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbwz7\" (UniqueName: \"kubernetes.io/projected/f69da577-0161-4756-94be-12ee983945bd-kube-api-access-vbwz7\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.397980 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.475068 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-df5bp" event={"ID":"f69da577-0161-4756-94be-12ee983945bd","Type":"ContainerDied","Data":"91e3d42624467e2f64f8614ac8a9f080d315ea947f5544be62385500251f5a77"} Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.475430 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91e3d42624467e2f64f8614ac8a9f080d315ea947f5544be62385500251f5a77" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.475511 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df5bp" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.479046 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5z5q5" event={"ID":"6a4bdf24-ed65-4aa7-acd9-922bd48a1b88","Type":"ContainerDied","Data":"82101d0ef334954794dec5caf10128501d7fd3002adf9b717e9ddb1767fa225e"} Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.479086 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82101d0ef334954794dec5caf10128501d7fd3002adf9b717e9ddb1767fa225e" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.479173 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z5q5" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.483630 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z7bqr" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.483816 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z7bqr" event={"ID":"67017489-fd71-4e43-99c6-7918bc9f41f5","Type":"ContainerDied","Data":"f35b2148706eddd5194cd6b276978f42649b76ef06976b74a80cf8e0de6c7ea9"} Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.483948 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f35b2148706eddd5194cd6b276978f42649b76ef06976b74a80cf8e0de6c7ea9" Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.494066 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerStarted","Data":"cc501e14690589f12d0b4a79cc4ad29996be4a52a66d9b720f3319d9f8506044"} Oct 08 15:35:47 crc kubenswrapper[4945]: I1008 15:35:47.851689 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:35:48 crc kubenswrapper[4945]: I1008 15:35:48.038246 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e10c8660-ea2f-483c-afbe-7f7e6287968c" path="/var/lib/kubelet/pods/e10c8660-ea2f-483c-afbe-7f7e6287968c/volumes" Oct 08 15:35:48 crc kubenswrapper[4945]: I1008 15:35:48.524358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerStarted","Data":"1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762"} Oct 08 15:35:48 crc kubenswrapper[4945]: I1008 15:35:48.525895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerStarted","Data":"70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28"} Oct 08 15:35:49 crc kubenswrapper[4945]: I1008 15:35:49.184473 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:35:49 crc kubenswrapper[4945]: I1008 15:35:49.184726 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:35:49 crc kubenswrapper[4945]: I1008 15:35:49.535880 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerStarted","Data":"bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a"} Oct 08 15:35:50 crc kubenswrapper[4945]: I1008 15:35:50.549542 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerStarted","Data":"c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33"} Oct 08 15:35:50 crc kubenswrapper[4945]: I1008 15:35:50.549716 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-central-agent" containerID="cri-o://70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28" gracePeriod=30 Oct 08 15:35:50 crc kubenswrapper[4945]: I1008 15:35:50.549992 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="proxy-httpd" containerID="cri-o://c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33" gracePeriod=30 Oct 08 15:35:50 crc kubenswrapper[4945]: I1008 15:35:50.550022 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-notification-agent" containerID="cri-o://1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762" gracePeriod=30 Oct 08 15:35:50 crc kubenswrapper[4945]: I1008 15:35:50.550083 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="sg-core" containerID="cri-o://bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a" gracePeriod=30 Oct 08 15:35:50 crc kubenswrapper[4945]: I1008 15:35:50.550148 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:35:50 crc kubenswrapper[4945]: I1008 15:35:50.578251 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.012764692 podStartE2EDuration="4.578230852s" podCreationTimestamp="2025-10-08 15:35:46 +0000 UTC" firstStartedPulling="2025-10-08 15:35:47.408212614 +0000 UTC m=+1356.762127535" lastFinishedPulling="2025-10-08 15:35:49.973678784 +0000 UTC m=+1359.327593695" observedRunningTime="2025-10-08 15:35:50.572379535 +0000 UTC m=+1359.926294456" watchObservedRunningTime="2025-10-08 15:35:50.578230852 +0000 UTC m=+1359.932145753" Oct 08 15:35:51 crc kubenswrapper[4945]: I1008 15:35:51.564630 4945 generic.go:334] "Generic (PLEG): container finished" podID="405d1d9b-075c-445a-9463-0d79a97338a0" containerID="c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33" exitCode=0 Oct 08 15:35:51 crc kubenswrapper[4945]: I1008 15:35:51.565065 4945 generic.go:334] "Generic (PLEG): container finished" podID="405d1d9b-075c-445a-9463-0d79a97338a0" containerID="bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a" exitCode=2 Oct 08 15:35:51 crc kubenswrapper[4945]: I1008 15:35:51.565080 4945 generic.go:334] "Generic (PLEG): container finished" podID="405d1d9b-075c-445a-9463-0d79a97338a0" containerID="1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762" exitCode=0 Oct 08 15:35:51 crc kubenswrapper[4945]: I1008 15:35:51.564688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerDied","Data":"c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33"} Oct 08 15:35:51 crc kubenswrapper[4945]: I1008 15:35:51.565146 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerDied","Data":"bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a"} Oct 08 15:35:51 crc kubenswrapper[4945]: I1008 15:35:51.565165 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerDied","Data":"1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762"} Oct 08 15:35:52 crc kubenswrapper[4945]: I1008 15:35:52.160140 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:52 crc kubenswrapper[4945]: I1008 15:35:52.160580 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-log" containerID="cri-o://e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8" gracePeriod=30 Oct 08 15:35:52 crc kubenswrapper[4945]: I1008 15:35:52.160705 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-httpd" containerID="cri-o://02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465" gracePeriod=30 Oct 08 15:35:52 crc kubenswrapper[4945]: I1008 15:35:52.170593 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.186:9292/healthcheck\": EOF" Oct 08 15:35:52 crc kubenswrapper[4945]: I1008 15:35:52.176122 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.186:9292/healthcheck\": EOF" Oct 08 15:35:52 crc kubenswrapper[4945]: I1008 15:35:52.579671 4945 generic.go:334] "Generic (PLEG): container finished" podID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerID="e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8" exitCode=143 Oct 08 15:35:52 crc kubenswrapper[4945]: I1008 15:35:52.579762 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41","Type":"ContainerDied","Data":"e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8"} Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.086690 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.087128 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-httpd" containerID="cri-o://c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77" gracePeriod=30 Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.087283 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-log" containerID="cri-o://a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2" gracePeriod=30 Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.096738 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.187:9292/healthcheck\": EOF" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.098153 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.187:9292/healthcheck\": read tcp 10.217.0.2:49080->10.217.0.187:9292: read: connection reset by peer" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.355463 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3781-account-create-qxkhh"] Oct 08 15:35:53 crc kubenswrapper[4945]: E1008 15:35:53.355907 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4bdf24-ed65-4aa7-acd9-922bd48a1b88" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.355923 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4bdf24-ed65-4aa7-acd9-922bd48a1b88" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: E1008 15:35:53.355952 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69da577-0161-4756-94be-12ee983945bd" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.355962 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69da577-0161-4756-94be-12ee983945bd" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: E1008 15:35:53.355985 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67017489-fd71-4e43-99c6-7918bc9f41f5" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.355992 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="67017489-fd71-4e43-99c6-7918bc9f41f5" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.356788 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a4bdf24-ed65-4aa7-acd9-922bd48a1b88" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.356813 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="67017489-fd71-4e43-99c6-7918bc9f41f5" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.356927 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69da577-0161-4756-94be-12ee983945bd" containerName="mariadb-database-create" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.357905 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3781-account-create-qxkhh" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.359894 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.371855 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3781-account-create-qxkhh"] Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.432430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/d55dc803-9da2-4af4-b64a-6d9b40b2fa06-kube-api-access-lxr2j\") pod \"nova-api-3781-account-create-qxkhh\" (UID: \"d55dc803-9da2-4af4-b64a-6d9b40b2fa06\") " pod="openstack/nova-api-3781-account-create-qxkhh" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.534344 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/d55dc803-9da2-4af4-b64a-6d9b40b2fa06-kube-api-access-lxr2j\") pod \"nova-api-3781-account-create-qxkhh\" (UID: \"d55dc803-9da2-4af4-b64a-6d9b40b2fa06\") " pod="openstack/nova-api-3781-account-create-qxkhh" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.555387 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-95e9-account-create-z2krj"] Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.556816 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95e9-account-create-z2krj" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.559963 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/d55dc803-9da2-4af4-b64a-6d9b40b2fa06-kube-api-access-lxr2j\") pod \"nova-api-3781-account-create-qxkhh\" (UID: \"d55dc803-9da2-4af4-b64a-6d9b40b2fa06\") " pod="openstack/nova-api-3781-account-create-qxkhh" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.560275 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.569269 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-95e9-account-create-z2krj"] Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.598220 4945 generic.go:334] "Generic (PLEG): container finished" podID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerID="a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2" exitCode=143 Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.598269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be98f960-821f-4340-9d7c-6cb26b2b330f","Type":"ContainerDied","Data":"a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2"} Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.636216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnvmw\" (UniqueName: \"kubernetes.io/projected/5c284362-1210-4291-b285-ec21e6c862e0-kube-api-access-mnvmw\") pod \"nova-cell0-95e9-account-create-z2krj\" (UID: \"5c284362-1210-4291-b285-ec21e6c862e0\") " pod="openstack/nova-cell0-95e9-account-create-z2krj" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.679174 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3781-account-create-qxkhh" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.737845 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnvmw\" (UniqueName: \"kubernetes.io/projected/5c284362-1210-4291-b285-ec21e6c862e0-kube-api-access-mnvmw\") pod \"nova-cell0-95e9-account-create-z2krj\" (UID: \"5c284362-1210-4291-b285-ec21e6c862e0\") " pod="openstack/nova-cell0-95e9-account-create-z2krj" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.825456 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f303-account-create-qwm6s"] Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.848952 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f303-account-create-qwm6s"] Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.849104 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f303-account-create-qwm6s" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.846962 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnvmw\" (UniqueName: \"kubernetes.io/projected/5c284362-1210-4291-b285-ec21e6c862e0-kube-api-access-mnvmw\") pod \"nova-cell0-95e9-account-create-z2krj\" (UID: \"5c284362-1210-4291-b285-ec21e6c862e0\") " pod="openstack/nova-cell0-95e9-account-create-z2krj" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.863499 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.942951 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j2sl\" (UniqueName: \"kubernetes.io/projected/5deefdab-a29a-4b5a-ad0d-08eccea748d1-kube-api-access-7j2sl\") pod \"nova-cell1-f303-account-create-qwm6s\" (UID: \"5deefdab-a29a-4b5a-ad0d-08eccea748d1\") " pod="openstack/nova-cell1-f303-account-create-qwm6s" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.966935 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95e9-account-create-z2krj" Oct 08 15:35:53 crc kubenswrapper[4945]: I1008 15:35:53.974308 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.027905 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:35:54 crc kubenswrapper[4945]: E1008 15:35:54.028370 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.045020 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j2sl\" (UniqueName: \"kubernetes.io/projected/5deefdab-a29a-4b5a-ad0d-08eccea748d1-kube-api-access-7j2sl\") pod \"nova-cell1-f303-account-create-qwm6s\" (UID: \"5deefdab-a29a-4b5a-ad0d-08eccea748d1\") " pod="openstack/nova-cell1-f303-account-create-qwm6s" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.118708 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j2sl\" (UniqueName: \"kubernetes.io/projected/5deefdab-a29a-4b5a-ad0d-08eccea748d1-kube-api-access-7j2sl\") pod \"nova-cell1-f303-account-create-qwm6s\" (UID: \"5deefdab-a29a-4b5a-ad0d-08eccea748d1\") " pod="openstack/nova-cell1-f303-account-create-qwm6s" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.250917 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f303-account-create-qwm6s" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.446861 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.474598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-scripts\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.474665 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.474746 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs22t\" (UniqueName: \"kubernetes.io/projected/be98f960-821f-4340-9d7c-6cb26b2b330f-kube-api-access-cs22t\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.474872 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-config-data\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.474952 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-combined-ca-bundle\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.475021 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-httpd-run\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.475067 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-internal-tls-certs\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.475137 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-logs\") pod \"be98f960-821f-4340-9d7c-6cb26b2b330f\" (UID: \"be98f960-821f-4340-9d7c-6cb26b2b330f\") " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.476696 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.480905 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-logs" (OuterVolumeSpecName: "logs") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.486341 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.498483 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-scripts" (OuterVolumeSpecName: "scripts") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.500454 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3781-account-create-qxkhh"] Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.509632 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be98f960-821f-4340-9d7c-6cb26b2b330f-kube-api-access-cs22t" (OuterVolumeSpecName: "kube-api-access-cs22t") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "kube-api-access-cs22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: W1008 15:35:54.550778 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd55dc803_9da2_4af4_b64a_6d9b40b2fa06.slice/crio-214cd316f52b20c087f005cf23e720c3888d29f9cefe140cfd7adf0f996a15bf WatchSource:0}: Error finding container 214cd316f52b20c087f005cf23e720c3888d29f9cefe140cfd7adf0f996a15bf: Status 404 returned error can't find the container with id 214cd316f52b20c087f005cf23e720c3888d29f9cefe140cfd7adf0f996a15bf Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.580904 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.580959 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be98f960-821f-4340-9d7c-6cb26b2b330f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.580970 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.581021 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.581035 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs22t\" (UniqueName: \"kubernetes.io/projected/be98f960-821f-4340-9d7c-6cb26b2b330f-kube-api-access-cs22t\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.585276 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.587348 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.617933 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.648886 4945 generic.go:334] "Generic (PLEG): container finished" podID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerID="c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77" exitCode=0 Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.648969 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be98f960-821f-4340-9d7c-6cb26b2b330f","Type":"ContainerDied","Data":"c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77"} Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.649002 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be98f960-821f-4340-9d7c-6cb26b2b330f","Type":"ContainerDied","Data":"282b8867efb95f43a1e609a575bac3ee062b1a10fb2daab1a5915996ef1f7b4d"} Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.649019 4945 scope.go:117] "RemoveContainer" containerID="c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.649173 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.653969 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3781-account-create-qxkhh" event={"ID":"d55dc803-9da2-4af4-b64a-6d9b40b2fa06","Type":"ContainerStarted","Data":"214cd316f52b20c087f005cf23e720c3888d29f9cefe140cfd7adf0f996a15bf"} Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.683172 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.683201 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.683209 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.695422 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-config-data" (OuterVolumeSpecName: "config-data") pod "be98f960-821f-4340-9d7c-6cb26b2b330f" (UID: "be98f960-821f-4340-9d7c-6cb26b2b330f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.786412 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be98f960-821f-4340-9d7c-6cb26b2b330f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.794307 4945 scope.go:117] "RemoveContainer" containerID="a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.818029 4945 scope.go:117] "RemoveContainer" containerID="c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77" Oct 08 15:35:54 crc kubenswrapper[4945]: E1008 15:35:54.818516 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77\": container with ID starting with c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77 not found: ID does not exist" containerID="c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.818573 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77"} err="failed to get container status \"c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77\": rpc error: code = NotFound desc = could not find container \"c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77\": container with ID starting with c3337e3ad2b6ef49fe0bebcb9c9e94e042dd29c429f500439354e374677a3c77 not found: ID does not exist" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.818600 4945 scope.go:117] "RemoveContainer" containerID="a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2" Oct 08 15:35:54 crc kubenswrapper[4945]: E1008 15:35:54.819038 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2\": container with ID starting with a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2 not found: ID does not exist" containerID="a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.819078 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2"} err="failed to get container status \"a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2\": rpc error: code = NotFound desc = could not find container \"a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2\": container with ID starting with a1dcc8c44028901b2b10ca5209604fea0aab2c86ad32e9686ed25ff65834aee2 not found: ID does not exist" Oct 08 15:35:54 crc kubenswrapper[4945]: I1008 15:35:54.860663 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-95e9-account-create-z2krj"] Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.015460 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.028011 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.047100 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:55 crc kubenswrapper[4945]: E1008 15:35:55.049190 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-log" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.049210 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-log" Oct 08 15:35:55 crc kubenswrapper[4945]: E1008 15:35:55.049240 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-httpd" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.053563 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-httpd" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.053841 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-log" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.053867 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" containerName="glance-httpd" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.055151 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.059946 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.061024 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.078151 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093202 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj8sm\" (UniqueName: \"kubernetes.io/projected/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-kube-api-access-hj8sm\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093300 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093355 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093399 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093429 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093525 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.093544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.114402 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f303-account-create-qwm6s"] Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.195827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.195869 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.195914 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj8sm\" (UniqueName: \"kubernetes.io/projected/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-kube-api-access-hj8sm\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.195971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.195991 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.196018 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.196045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.196072 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.198493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.198539 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.198742 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.203008 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.203227 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.204414 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.206656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.214433 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj8sm\" (UniqueName: \"kubernetes.io/projected/f2f0a0ac-d149-4fc1-878a-7e21188cd4ee-kube-api-access-hj8sm\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.240828 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee\") " pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.378145 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.682133 4945 generic.go:334] "Generic (PLEG): container finished" podID="5deefdab-a29a-4b5a-ad0d-08eccea748d1" containerID="fbb92a6973d3c5dbb73613c7d37fc0e8876a26d3027f5d940253ed02c7adb22d" exitCode=0 Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.683050 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f303-account-create-qwm6s" event={"ID":"5deefdab-a29a-4b5a-ad0d-08eccea748d1","Type":"ContainerDied","Data":"fbb92a6973d3c5dbb73613c7d37fc0e8876a26d3027f5d940253ed02c7adb22d"} Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.683092 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f303-account-create-qwm6s" event={"ID":"5deefdab-a29a-4b5a-ad0d-08eccea748d1","Type":"ContainerStarted","Data":"b29d805201ebabadb534d2df21406637687f19b42a73641f26fa67f4f2a83221"} Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.709536 4945 generic.go:334] "Generic (PLEG): container finished" podID="5c284362-1210-4291-b285-ec21e6c862e0" containerID="811ed07885f90146efc50daa22a8f32f58b52c6d8d2443fb9766122e2be554c1" exitCode=0 Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.709614 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-95e9-account-create-z2krj" event={"ID":"5c284362-1210-4291-b285-ec21e6c862e0","Type":"ContainerDied","Data":"811ed07885f90146efc50daa22a8f32f58b52c6d8d2443fb9766122e2be554c1"} Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.709648 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-95e9-account-create-z2krj" event={"ID":"5c284362-1210-4291-b285-ec21e6c862e0","Type":"ContainerStarted","Data":"d9dc53b31885cfda1bad563d4bc11a51ec555295f8d792acd9705894ad0514bb"} Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.723308 4945 generic.go:334] "Generic (PLEG): container finished" podID="d55dc803-9da2-4af4-b64a-6d9b40b2fa06" containerID="de1bf0ed5f75e054c3099e66e65dbc1dd77273c439f48b8cb404a009e1613a21" exitCode=0 Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.723394 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3781-account-create-qxkhh" event={"ID":"d55dc803-9da2-4af4-b64a-6d9b40b2fa06","Type":"ContainerDied","Data":"de1bf0ed5f75e054c3099e66e65dbc1dd77273c439f48b8cb404a009e1613a21"} Oct 08 15:35:55 crc kubenswrapper[4945]: I1008 15:35:55.913890 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 15:35:56 crc kubenswrapper[4945]: I1008 15:35:56.038187 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be98f960-821f-4340-9d7c-6cb26b2b330f" path="/var/lib/kubelet/pods/be98f960-821f-4340-9d7c-6cb26b2b330f/volumes" Oct 08 15:35:56 crc kubenswrapper[4945]: I1008 15:35:56.748466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee","Type":"ContainerStarted","Data":"abd925df58fbdb3bb11e6a56b7550d0a807ec11d9d17ad759a36455d9bcae7ae"} Oct 08 15:35:56 crc kubenswrapper[4945]: I1008 15:35:56.748776 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee","Type":"ContainerStarted","Data":"4ec0f95e1d620509ae33999ad634de280a56739becd1a28dfaf96314ef551366"} Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.037960 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5988468485-ctpxh" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.121256 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-779fb94656-nstv6"] Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.121472 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-779fb94656-nstv6" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-api" containerID="cri-o://43d7c5ffa4eeae66140beeb5081c713ba6729a331de1f70c55ac886e9dff4589" gracePeriod=30 Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.121865 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-779fb94656-nstv6" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-httpd" containerID="cri-o://745d9a9caca513603cff2b68e8fe7563b8ed2a7ab5517886916c060ca002e593" gracePeriod=30 Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.339958 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f303-account-create-qwm6s" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.448201 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7j2sl\" (UniqueName: \"kubernetes.io/projected/5deefdab-a29a-4b5a-ad0d-08eccea748d1-kube-api-access-7j2sl\") pod \"5deefdab-a29a-4b5a-ad0d-08eccea748d1\" (UID: \"5deefdab-a29a-4b5a-ad0d-08eccea748d1\") " Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.458005 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5deefdab-a29a-4b5a-ad0d-08eccea748d1-kube-api-access-7j2sl" (OuterVolumeSpecName: "kube-api-access-7j2sl") pod "5deefdab-a29a-4b5a-ad0d-08eccea748d1" (UID: "5deefdab-a29a-4b5a-ad0d-08eccea748d1"). InnerVolumeSpecName "kube-api-access-7j2sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.470898 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95e9-account-create-z2krj" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.477619 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3781-account-create-qxkhh" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.549865 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/d55dc803-9da2-4af4-b64a-6d9b40b2fa06-kube-api-access-lxr2j\") pod \"d55dc803-9da2-4af4-b64a-6d9b40b2fa06\" (UID: \"d55dc803-9da2-4af4-b64a-6d9b40b2fa06\") " Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.549935 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnvmw\" (UniqueName: \"kubernetes.io/projected/5c284362-1210-4291-b285-ec21e6c862e0-kube-api-access-mnvmw\") pod \"5c284362-1210-4291-b285-ec21e6c862e0\" (UID: \"5c284362-1210-4291-b285-ec21e6c862e0\") " Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.550391 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7j2sl\" (UniqueName: \"kubernetes.io/projected/5deefdab-a29a-4b5a-ad0d-08eccea748d1-kube-api-access-7j2sl\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.554902 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d55dc803-9da2-4af4-b64a-6d9b40b2fa06-kube-api-access-lxr2j" (OuterVolumeSpecName: "kube-api-access-lxr2j") pod "d55dc803-9da2-4af4-b64a-6d9b40b2fa06" (UID: "d55dc803-9da2-4af4-b64a-6d9b40b2fa06"). InnerVolumeSpecName "kube-api-access-lxr2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.558220 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c284362-1210-4291-b285-ec21e6c862e0-kube-api-access-mnvmw" (OuterVolumeSpecName: "kube-api-access-mnvmw") pod "5c284362-1210-4291-b285-ec21e6c862e0" (UID: "5c284362-1210-4291-b285-ec21e6c862e0"). InnerVolumeSpecName "kube-api-access-mnvmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.652218 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/d55dc803-9da2-4af4-b64a-6d9b40b2fa06-kube-api-access-lxr2j\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.652252 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnvmw\" (UniqueName: \"kubernetes.io/projected/5c284362-1210-4291-b285-ec21e6c862e0-kube-api-access-mnvmw\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.757794 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f303-account-create-qwm6s" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.757815 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f303-account-create-qwm6s" event={"ID":"5deefdab-a29a-4b5a-ad0d-08eccea748d1","Type":"ContainerDied","Data":"b29d805201ebabadb534d2df21406637687f19b42a73641f26fa67f4f2a83221"} Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.758143 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b29d805201ebabadb534d2df21406637687f19b42a73641f26fa67f4f2a83221" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.759947 4945 generic.go:334] "Generic (PLEG): container finished" podID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerID="745d9a9caca513603cff2b68e8fe7563b8ed2a7ab5517886916c060ca002e593" exitCode=0 Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.759989 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-779fb94656-nstv6" event={"ID":"48be9df6-e6b0-4345-8237-ea4a52cd03f7","Type":"ContainerDied","Data":"745d9a9caca513603cff2b68e8fe7563b8ed2a7ab5517886916c060ca002e593"} Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.761386 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f2f0a0ac-d149-4fc1-878a-7e21188cd4ee","Type":"ContainerStarted","Data":"525b2b26ee6ae76dc6cd6de4130c02b267c287cafe1d4b1afce97aabe80daa76"} Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.763867 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-95e9-account-create-z2krj" event={"ID":"5c284362-1210-4291-b285-ec21e6c862e0","Type":"ContainerDied","Data":"d9dc53b31885cfda1bad563d4bc11a51ec555295f8d792acd9705894ad0514bb"} Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.763889 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9dc53b31885cfda1bad563d4bc11a51ec555295f8d792acd9705894ad0514bb" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.763933 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95e9-account-create-z2krj" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.773235 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3781-account-create-qxkhh" event={"ID":"d55dc803-9da2-4af4-b64a-6d9b40b2fa06","Type":"ContainerDied","Data":"214cd316f52b20c087f005cf23e720c3888d29f9cefe140cfd7adf0f996a15bf"} Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.773344 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="214cd316f52b20c087f005cf23e720c3888d29f9cefe140cfd7adf0f996a15bf" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.773316 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3781-account-create-qxkhh" Oct 08 15:35:57 crc kubenswrapper[4945]: I1008 15:35:57.811165 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.81114534 podStartE2EDuration="2.81114534s" podCreationTimestamp="2025-10-08 15:35:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:35:57.802428581 +0000 UTC m=+1367.156343482" watchObservedRunningTime="2025-10-08 15:35:57.81114534 +0000 UTC m=+1367.165060261" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.728597 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.794770 4945 generic.go:334] "Generic (PLEG): container finished" podID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerID="02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465" exitCode=0 Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.794856 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.794897 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41","Type":"ContainerDied","Data":"02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465"} Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.794928 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41","Type":"ContainerDied","Data":"251a36aa8e95d984750d1c7613aef7dc9d260fd99c89a358205b6e7d5b254081"} Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.794944 4945 scope.go:117] "RemoveContainer" containerID="02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834068 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jh7mz"] Oct 08 15:35:58 crc kubenswrapper[4945]: E1008 15:35:58.834498 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-httpd" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834510 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-httpd" Oct 08 15:35:58 crc kubenswrapper[4945]: E1008 15:35:58.834543 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c284362-1210-4291-b285-ec21e6c862e0" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834549 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c284362-1210-4291-b285-ec21e6c862e0" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: E1008 15:35:58.834563 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d55dc803-9da2-4af4-b64a-6d9b40b2fa06" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834569 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d55dc803-9da2-4af4-b64a-6d9b40b2fa06" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: E1008 15:35:58.834577 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5deefdab-a29a-4b5a-ad0d-08eccea748d1" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834584 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5deefdab-a29a-4b5a-ad0d-08eccea748d1" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: E1008 15:35:58.834593 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-log" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834599 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-log" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834827 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c284362-1210-4291-b285-ec21e6c862e0" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834844 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d55dc803-9da2-4af4-b64a-6d9b40b2fa06" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834859 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-httpd" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834871 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5deefdab-a29a-4b5a-ad0d-08eccea748d1" containerName="mariadb-account-create" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.834883 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" containerName="glance-log" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.835635 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.839771 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.839993 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r5k2t" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.840066 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.845079 4945 scope.go:117] "RemoveContainer" containerID="e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.846541 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jh7mz"] Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.878321 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-combined-ca-bundle\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.878457 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.878575 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-config-data\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.878623 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-logs\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.880063 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhcpg\" (UniqueName: \"kubernetes.io/projected/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-kube-api-access-jhcpg\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.880142 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-public-tls-certs\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.880443 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-logs" (OuterVolumeSpecName: "logs") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.881559 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-httpd-run\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.882240 4945 scope.go:117] "RemoveContainer" containerID="02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.882720 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.883643 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: E1008 15:35:58.883651 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465\": container with ID starting with 02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465 not found: ID does not exist" containerID="02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.883677 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465"} err="failed to get container status \"02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465\": rpc error: code = NotFound desc = could not find container \"02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465\": container with ID starting with 02c30240398bf9bf113c31554d7de73ab46c81bd88e3e06832a65ed007a7d465 not found: ID does not exist" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.883701 4945 scope.go:117] "RemoveContainer" containerID="e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.884760 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-scripts\") pod \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\" (UID: \"2bf19d0f-a799-43f5-b39c-c9d98fdf0e41\") " Oct 08 15:35:58 crc kubenswrapper[4945]: E1008 15:35:58.887261 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8\": container with ID starting with e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8 not found: ID does not exist" containerID="e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.887297 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8"} err="failed to get container status \"e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8\": rpc error: code = NotFound desc = could not find container \"e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8\": container with ID starting with e75c763d4fd93e50d2bacbb610b7f06ca5d5123a308b246df4c2089e540b3fc8 not found: ID does not exist" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.888360 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.888415 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.888605 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.910468 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-scripts" (OuterVolumeSpecName: "scripts") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.925289 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-kube-api-access-jhcpg" (OuterVolumeSpecName: "kube-api-access-jhcpg") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "kube-api-access-jhcpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.929772 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.955122 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.965422 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-config-data" (OuterVolumeSpecName: "config-data") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.969049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" (UID: "2bf19d0f-a799-43f5-b39c-c9d98fdf0e41"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.990217 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9xkz\" (UniqueName: \"kubernetes.io/projected/59d1ed3b-11c9-4deb-9947-605022b6592d-kube-api-access-l9xkz\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.990607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-scripts\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991595 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991672 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-config-data\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991757 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhcpg\" (UniqueName: \"kubernetes.io/projected/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-kube-api-access-jhcpg\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991777 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991798 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991812 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991825 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:58 crc kubenswrapper[4945]: I1008 15:35:58.991839 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.100027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.100126 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-config-data\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.100202 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9xkz\" (UniqueName: \"kubernetes.io/projected/59d1ed3b-11c9-4deb-9947-605022b6592d-kube-api-access-l9xkz\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.100227 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-scripts\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.113517 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-config-data\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.116520 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.120451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-scripts\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.130400 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9xkz\" (UniqueName: \"kubernetes.io/projected/59d1ed3b-11c9-4deb-9947-605022b6592d-kube-api-access-l9xkz\") pod \"nova-cell0-conductor-db-sync-jh7mz\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.164526 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.165225 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.183750 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.204034 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.206298 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.209308 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.210554 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.231378 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304469 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304524 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-scripts\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304550 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304596 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-config-data\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304612 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304647 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc16ff2a-fe01-42bf-8b16-e4ba10315929-logs\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304696 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc16ff2a-fe01-42bf-8b16-e4ba10315929-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.304760 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfhz9\" (UniqueName: \"kubernetes.io/projected/bc16ff2a-fe01-42bf-8b16-e4ba10315929-kube-api-access-wfhz9\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406247 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc16ff2a-fe01-42bf-8b16-e4ba10315929-logs\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406541 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc16ff2a-fe01-42bf-8b16-e4ba10315929-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406613 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfhz9\" (UniqueName: \"kubernetes.io/projected/bc16ff2a-fe01-42bf-8b16-e4ba10315929-kube-api-access-wfhz9\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-scripts\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406698 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406721 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-config-data\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406737 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406746 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc16ff2a-fe01-42bf-8b16-e4ba10315929-logs\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.406935 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc16ff2a-fe01-42bf-8b16-e4ba10315929-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.407049 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.410646 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-scripts\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.413765 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.417312 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-config-data\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.418403 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc16ff2a-fe01-42bf-8b16-e4ba10315929-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.425989 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfhz9\" (UniqueName: \"kubernetes.io/projected/bc16ff2a-fe01-42bf-8b16-e4ba10315929-kube-api-access-wfhz9\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.441601 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.480559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bc16ff2a-fe01-42bf-8b16-e4ba10315929\") " pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.535685 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.612899 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-run-httpd\") pod \"405d1d9b-075c-445a-9463-0d79a97338a0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.612999 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7lxn\" (UniqueName: \"kubernetes.io/projected/405d1d9b-075c-445a-9463-0d79a97338a0-kube-api-access-p7lxn\") pod \"405d1d9b-075c-445a-9463-0d79a97338a0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.613070 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-scripts\") pod \"405d1d9b-075c-445a-9463-0d79a97338a0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.613168 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-config-data\") pod \"405d1d9b-075c-445a-9463-0d79a97338a0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.613206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-combined-ca-bundle\") pod \"405d1d9b-075c-445a-9463-0d79a97338a0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.613225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-log-httpd\") pod \"405d1d9b-075c-445a-9463-0d79a97338a0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.613261 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-sg-core-conf-yaml\") pod \"405d1d9b-075c-445a-9463-0d79a97338a0\" (UID: \"405d1d9b-075c-445a-9463-0d79a97338a0\") " Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.613997 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "405d1d9b-075c-445a-9463-0d79a97338a0" (UID: "405d1d9b-075c-445a-9463-0d79a97338a0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.616927 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "405d1d9b-075c-445a-9463-0d79a97338a0" (UID: "405d1d9b-075c-445a-9463-0d79a97338a0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.617057 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405d1d9b-075c-445a-9463-0d79a97338a0-kube-api-access-p7lxn" (OuterVolumeSpecName: "kube-api-access-p7lxn") pod "405d1d9b-075c-445a-9463-0d79a97338a0" (UID: "405d1d9b-075c-445a-9463-0d79a97338a0"). InnerVolumeSpecName "kube-api-access-p7lxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.625883 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-scripts" (OuterVolumeSpecName: "scripts") pod "405d1d9b-075c-445a-9463-0d79a97338a0" (UID: "405d1d9b-075c-445a-9463-0d79a97338a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.665004 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "405d1d9b-075c-445a-9463-0d79a97338a0" (UID: "405d1d9b-075c-445a-9463-0d79a97338a0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.723622 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7lxn\" (UniqueName: \"kubernetes.io/projected/405d1d9b-075c-445a-9463-0d79a97338a0-kube-api-access-p7lxn\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.723660 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.723669 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.723677 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.723684 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/405d1d9b-075c-445a-9463-0d79a97338a0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.752025 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jh7mz"] Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.770102 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "405d1d9b-075c-445a-9463-0d79a97338a0" (UID: "405d1d9b-075c-445a-9463-0d79a97338a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.817368 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" event={"ID":"59d1ed3b-11c9-4deb-9947-605022b6592d","Type":"ContainerStarted","Data":"f363267d111614e9ed2f0a43c30e884b74e690bd370ceaeb5815ef2093665f61"} Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.830964 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.839701 4945 generic.go:334] "Generic (PLEG): container finished" podID="405d1d9b-075c-445a-9463-0d79a97338a0" containerID="70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28" exitCode=0 Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.839744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerDied","Data":"70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28"} Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.839771 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"405d1d9b-075c-445a-9463-0d79a97338a0","Type":"ContainerDied","Data":"cc501e14690589f12d0b4a79cc4ad29996be4a52a66d9b720f3319d9f8506044"} Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.839788 4945 scope.go:117] "RemoveContainer" containerID="c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.839903 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.867013 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-config-data" (OuterVolumeSpecName: "config-data") pod "405d1d9b-075c-445a-9463-0d79a97338a0" (UID: "405d1d9b-075c-445a-9463-0d79a97338a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.867208 4945 scope.go:117] "RemoveContainer" containerID="bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.917255 4945 scope.go:117] "RemoveContainer" containerID="1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.932455 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d1d9b-075c-445a-9463-0d79a97338a0-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.952556 4945 scope.go:117] "RemoveContainer" containerID="70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.990961 4945 scope.go:117] "RemoveContainer" containerID="c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33" Oct 08 15:35:59 crc kubenswrapper[4945]: E1008 15:35:59.993282 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33\": container with ID starting with c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33 not found: ID does not exist" containerID="c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.993334 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33"} err="failed to get container status \"c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33\": rpc error: code = NotFound desc = could not find container \"c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33\": container with ID starting with c5ae58095d7425ba15f52bee8be96b483c8a0ab12fe49c8d1c1580576fa09c33 not found: ID does not exist" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.993362 4945 scope.go:117] "RemoveContainer" containerID="bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a" Oct 08 15:35:59 crc kubenswrapper[4945]: E1008 15:35:59.994272 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a\": container with ID starting with bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a not found: ID does not exist" containerID="bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.994302 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a"} err="failed to get container status \"bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a\": rpc error: code = NotFound desc = could not find container \"bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a\": container with ID starting with bcf5e392b97be0ef32a5f293a27e890738e14f5a5b25e4d078cb59489936f84a not found: ID does not exist" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.994327 4945 scope.go:117] "RemoveContainer" containerID="1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762" Oct 08 15:35:59 crc kubenswrapper[4945]: E1008 15:35:59.994533 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762\": container with ID starting with 1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762 not found: ID does not exist" containerID="1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.994550 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762"} err="failed to get container status \"1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762\": rpc error: code = NotFound desc = could not find container \"1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762\": container with ID starting with 1dd196c604d10236a49a407ad674587cb845e27682010550c9455fdadf95e762 not found: ID does not exist" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.994563 4945 scope.go:117] "RemoveContainer" containerID="70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28" Oct 08 15:35:59 crc kubenswrapper[4945]: E1008 15:35:59.994786 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28\": container with ID starting with 70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28 not found: ID does not exist" containerID="70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28" Oct 08 15:35:59 crc kubenswrapper[4945]: I1008 15:35:59.994802 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28"} err="failed to get container status \"70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28\": rpc error: code = NotFound desc = could not find container \"70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28\": container with ID starting with 70295d3eff8b1958d3da95548e65d9caa2c58be940bca825b4c01125a8cbdf28 not found: ID does not exist" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.044313 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bf19d0f-a799-43f5-b39c-c9d98fdf0e41" path="/var/lib/kubelet/pods/2bf19d0f-a799-43f5-b39c-c9d98fdf0e41/volumes" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.090217 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.171541 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.187218 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.204721 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:00 crc kubenswrapper[4945]: E1008 15:36:00.206642 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="sg-core" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.206670 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="sg-core" Oct 08 15:36:00 crc kubenswrapper[4945]: E1008 15:36:00.206696 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="proxy-httpd" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.206704 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="proxy-httpd" Oct 08 15:36:00 crc kubenswrapper[4945]: E1008 15:36:00.206742 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-notification-agent" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.206751 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-notification-agent" Oct 08 15:36:00 crc kubenswrapper[4945]: E1008 15:36:00.206780 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-central-agent" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.206790 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-central-agent" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.207761 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="sg-core" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.207793 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-central-agent" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.207842 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="proxy-httpd" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.207863 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" containerName="ceilometer-notification-agent" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.218046 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.221812 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.221968 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.260803 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.345578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.345626 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-scripts\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.345644 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-run-httpd\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.345676 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-log-httpd\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.345754 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-config-data\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.346151 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.346188 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c9gc\" (UniqueName: \"kubernetes.io/projected/e681ef02-7b7d-40c7-9a6c-16d777364891-kube-api-access-5c9gc\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.448215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.449084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-scripts\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.449144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-run-httpd\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.449189 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-log-httpd\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.449241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-config-data\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.449337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.449367 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c9gc\" (UniqueName: \"kubernetes.io/projected/e681ef02-7b7d-40c7-9a6c-16d777364891-kube-api-access-5c9gc\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.450441 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-log-httpd\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.450687 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-run-httpd\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.455663 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-config-data\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.456098 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-scripts\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.458195 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.459678 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.466049 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c9gc\" (UniqueName: \"kubernetes.io/projected/e681ef02-7b7d-40c7-9a6c-16d777364891-kube-api-access-5c9gc\") pod \"ceilometer-0\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.552633 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.871394 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bc16ff2a-fe01-42bf-8b16-e4ba10315929","Type":"ContainerStarted","Data":"9102e74d62b68af87210fcf7e56abb45fb4cfcd0f21e48200f72df4ac587cffc"} Oct 08 15:36:00 crc kubenswrapper[4945]: I1008 15:36:00.871727 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bc16ff2a-fe01-42bf-8b16-e4ba10315929","Type":"ContainerStarted","Data":"0f63a771b5a5f7b791739e01aed95c8994b353e259e24badfc36dd7e6dc96ac8"} Oct 08 15:36:01 crc kubenswrapper[4945]: I1008 15:36:01.155482 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:01 crc kubenswrapper[4945]: W1008 15:36:01.167900 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode681ef02_7b7d_40c7_9a6c_16d777364891.slice/crio-1ad05614a57d989096a76edf13456f46825be5be77f57a1aa051fc289b69d7e1 WatchSource:0}: Error finding container 1ad05614a57d989096a76edf13456f46825be5be77f57a1aa051fc289b69d7e1: Status 404 returned error can't find the container with id 1ad05614a57d989096a76edf13456f46825be5be77f57a1aa051fc289b69d7e1 Oct 08 15:36:01 crc kubenswrapper[4945]: I1008 15:36:01.903665 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bc16ff2a-fe01-42bf-8b16-e4ba10315929","Type":"ContainerStarted","Data":"feebf8f38b1d78f36530d8c6132f27d0450ed322f4340c6021a7aada5f8ea02d"} Oct 08 15:36:01 crc kubenswrapper[4945]: I1008 15:36:01.907774 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerStarted","Data":"d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0"} Oct 08 15:36:01 crc kubenswrapper[4945]: I1008 15:36:01.907813 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerStarted","Data":"8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15"} Oct 08 15:36:01 crc kubenswrapper[4945]: I1008 15:36:01.907822 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerStarted","Data":"1ad05614a57d989096a76edf13456f46825be5be77f57a1aa051fc289b69d7e1"} Oct 08 15:36:01 crc kubenswrapper[4945]: I1008 15:36:01.927501 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.927485212 podStartE2EDuration="2.927485212s" podCreationTimestamp="2025-10-08 15:35:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:36:01.924487277 +0000 UTC m=+1371.278402168" watchObservedRunningTime="2025-10-08 15:36:01.927485212 +0000 UTC m=+1371.281400103" Oct 08 15:36:02 crc kubenswrapper[4945]: I1008 15:36:02.039699 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405d1d9b-075c-445a-9463-0d79a97338a0" path="/var/lib/kubelet/pods/405d1d9b-075c-445a-9463-0d79a97338a0/volumes" Oct 08 15:36:02 crc kubenswrapper[4945]: I1008 15:36:02.192754 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:02 crc kubenswrapper[4945]: I1008 15:36:02.193902 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:36:02 crc kubenswrapper[4945]: E1008 15:36:02.194205 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:36:02 crc kubenswrapper[4945]: I1008 15:36:02.197056 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:02 crc kubenswrapper[4945]: I1008 15:36:02.917736 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:36:02 crc kubenswrapper[4945]: E1008 15:36:02.917982 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.380720 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.381307 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.432591 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.436601 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.948210 4945 generic.go:334] "Generic (PLEG): container finished" podID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerID="43d7c5ffa4eeae66140beeb5081c713ba6729a331de1f70c55ac886e9dff4589" exitCode=0 Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.948306 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-779fb94656-nstv6" event={"ID":"48be9df6-e6b0-4345-8237-ea4a52cd03f7","Type":"ContainerDied","Data":"43d7c5ffa4eeae66140beeb5081c713ba6729a331de1f70c55ac886e9dff4589"} Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.948528 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:05 crc kubenswrapper[4945]: I1008 15:36:05.948559 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:08 crc kubenswrapper[4945]: I1008 15:36:08.960566 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:08 crc kubenswrapper[4945]: I1008 15:36:08.961355 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:36:08 crc kubenswrapper[4945]: I1008 15:36:08.965731 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 15:36:09 crc kubenswrapper[4945]: I1008 15:36:09.536138 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 15:36:09 crc kubenswrapper[4945]: I1008 15:36:09.536479 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 15:36:09 crc kubenswrapper[4945]: I1008 15:36:09.591544 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 15:36:09 crc kubenswrapper[4945]: I1008 15:36:09.600967 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 15:36:09 crc kubenswrapper[4945]: I1008 15:36:09.995361 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 15:36:09 crc kubenswrapper[4945]: I1008 15:36:09.995405 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 15:36:14 crc kubenswrapper[4945]: I1008 15:36:14.024984 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:36:14 crc kubenswrapper[4945]: E1008 15:36:14.025655 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(c464851c-cb2a-4b05-aac2-b3ee6f7c9182)\"" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" Oct 08 15:36:16 crc kubenswrapper[4945]: E1008 15:36:16.572135 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest" Oct 08 15:36:16 crc kubenswrapper[4945]: E1008 15:36:16.572618 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.163:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest" Oct 08 15:36:16 crc kubenswrapper[4945]: E1008 15:36:16.572735 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:38.102.83.163:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9xkz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-jh7mz_openstack(59d1ed3b-11c9-4deb-9947-605022b6592d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 15:36:16 crc kubenswrapper[4945]: E1008 15:36:16.574283 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" podUID="59d1ed3b-11c9-4deb-9947-605022b6592d" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.624963 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.773611 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67fqp\" (UniqueName: \"kubernetes.io/projected/48be9df6-e6b0-4345-8237-ea4a52cd03f7-kube-api-access-67fqp\") pod \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.773683 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-ovndb-tls-certs\") pod \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.773794 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-config\") pod \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.773941 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-combined-ca-bundle\") pod \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.774036 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-httpd-config\") pod \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\" (UID: \"48be9df6-e6b0-4345-8237-ea4a52cd03f7\") " Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.779314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "48be9df6-e6b0-4345-8237-ea4a52cd03f7" (UID: "48be9df6-e6b0-4345-8237-ea4a52cd03f7"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.798644 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48be9df6-e6b0-4345-8237-ea4a52cd03f7-kube-api-access-67fqp" (OuterVolumeSpecName: "kube-api-access-67fqp") pod "48be9df6-e6b0-4345-8237-ea4a52cd03f7" (UID: "48be9df6-e6b0-4345-8237-ea4a52cd03f7"). InnerVolumeSpecName "kube-api-access-67fqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.829642 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-config" (OuterVolumeSpecName: "config") pod "48be9df6-e6b0-4345-8237-ea4a52cd03f7" (UID: "48be9df6-e6b0-4345-8237-ea4a52cd03f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.862639 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48be9df6-e6b0-4345-8237-ea4a52cd03f7" (UID: "48be9df6-e6b0-4345-8237-ea4a52cd03f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.877433 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.877470 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.877481 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67fqp\" (UniqueName: \"kubernetes.io/projected/48be9df6-e6b0-4345-8237-ea4a52cd03f7-kube-api-access-67fqp\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.877495 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.882227 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "48be9df6-e6b0-4345-8237-ea4a52cd03f7" (UID: "48be9df6-e6b0-4345-8237-ea4a52cd03f7"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:16 crc kubenswrapper[4945]: I1008 15:36:16.979601 4945 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/48be9df6-e6b0-4345-8237-ea4a52cd03f7-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.069034 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerStarted","Data":"ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242"} Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.071203 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-779fb94656-nstv6" event={"ID":"48be9df6-e6b0-4345-8237-ea4a52cd03f7","Type":"ContainerDied","Data":"d3790c0a371b32b986f41780b0d6e6e4d8e34cf865402dad003dcc6dab224012"} Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.071272 4945 scope.go:117] "RemoveContainer" containerID="745d9a9caca513603cff2b68e8fe7563b8ed2a7ab5517886916c060ca002e593" Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.071225 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-779fb94656-nstv6" Oct 08 15:36:17 crc kubenswrapper[4945]: E1008 15:36:17.072877 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" podUID="59d1ed3b-11c9-4deb-9947-605022b6592d" Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.090814 4945 scope.go:117] "RemoveContainer" containerID="43d7c5ffa4eeae66140beeb5081c713ba6729a331de1f70c55ac886e9dff4589" Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.115359 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-779fb94656-nstv6"] Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.130086 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-779fb94656-nstv6"] Oct 08 15:36:17 crc kubenswrapper[4945]: I1008 15:36:17.141354 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:18 crc kubenswrapper[4945]: I1008 15:36:18.037993 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" path="/var/lib/kubelet/pods/48be9df6-e6b0-4345-8237-ea4a52cd03f7/volumes" Oct 08 15:36:18 crc kubenswrapper[4945]: I1008 15:36:18.128216 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 15:36:18 crc kubenswrapper[4945]: I1008 15:36:18.128633 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 15:36:18 crc kubenswrapper[4945]: I1008 15:36:18.174818 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.095769 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerStarted","Data":"c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27"} Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.096240 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.096073 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="sg-core" containerID="cri-o://ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242" gracePeriod=30 Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.096046 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="proxy-httpd" containerID="cri-o://c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27" gracePeriod=30 Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.096159 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-notification-agent" containerID="cri-o://d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0" gracePeriod=30 Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.096079 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-central-agent" containerID="cri-o://8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15" gracePeriod=30 Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.131233 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.030185932 podStartE2EDuration="19.13121328s" podCreationTimestamp="2025-10-08 15:36:00 +0000 UTC" firstStartedPulling="2025-10-08 15:36:01.188850735 +0000 UTC m=+1370.542765646" lastFinishedPulling="2025-10-08 15:36:18.289878093 +0000 UTC m=+1387.643792994" observedRunningTime="2025-10-08 15:36:19.120793598 +0000 UTC m=+1388.474708519" watchObservedRunningTime="2025-10-08 15:36:19.13121328 +0000 UTC m=+1388.485128181" Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.183793 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:36:19 crc kubenswrapper[4945]: I1008 15:36:19.183850 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:36:20 crc kubenswrapper[4945]: I1008 15:36:20.107228 4945 generic.go:334] "Generic (PLEG): container finished" podID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerID="c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27" exitCode=0 Oct 08 15:36:20 crc kubenswrapper[4945]: I1008 15:36:20.107628 4945 generic.go:334] "Generic (PLEG): container finished" podID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerID="ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242" exitCode=2 Oct 08 15:36:20 crc kubenswrapper[4945]: I1008 15:36:20.107341 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerDied","Data":"c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27"} Oct 08 15:36:20 crc kubenswrapper[4945]: I1008 15:36:20.107645 4945 generic.go:334] "Generic (PLEG): container finished" podID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerID="8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15" exitCode=0 Oct 08 15:36:20 crc kubenswrapper[4945]: I1008 15:36:20.107674 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerDied","Data":"ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242"} Oct 08 15:36:20 crc kubenswrapper[4945]: I1008 15:36:20.107693 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerDied","Data":"8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15"} Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.810214 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.899939 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-log-httpd\") pod \"e681ef02-7b7d-40c7-9a6c-16d777364891\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.900097 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c9gc\" (UniqueName: \"kubernetes.io/projected/e681ef02-7b7d-40c7-9a6c-16d777364891-kube-api-access-5c9gc\") pod \"e681ef02-7b7d-40c7-9a6c-16d777364891\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.900166 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-config-data\") pod \"e681ef02-7b7d-40c7-9a6c-16d777364891\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.900232 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-scripts\") pod \"e681ef02-7b7d-40c7-9a6c-16d777364891\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.900288 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-sg-core-conf-yaml\") pod \"e681ef02-7b7d-40c7-9a6c-16d777364891\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.900321 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-combined-ca-bundle\") pod \"e681ef02-7b7d-40c7-9a6c-16d777364891\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.900382 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-run-httpd\") pod \"e681ef02-7b7d-40c7-9a6c-16d777364891\" (UID: \"e681ef02-7b7d-40c7-9a6c-16d777364891\") " Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.900718 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e681ef02-7b7d-40c7-9a6c-16d777364891" (UID: "e681ef02-7b7d-40c7-9a6c-16d777364891"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.901203 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e681ef02-7b7d-40c7-9a6c-16d777364891" (UID: "e681ef02-7b7d-40c7-9a6c-16d777364891"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.906162 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-scripts" (OuterVolumeSpecName: "scripts") pod "e681ef02-7b7d-40c7-9a6c-16d777364891" (UID: "e681ef02-7b7d-40c7-9a6c-16d777364891"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.906718 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e681ef02-7b7d-40c7-9a6c-16d777364891-kube-api-access-5c9gc" (OuterVolumeSpecName: "kube-api-access-5c9gc") pod "e681ef02-7b7d-40c7-9a6c-16d777364891" (UID: "e681ef02-7b7d-40c7-9a6c-16d777364891"). InnerVolumeSpecName "kube-api-access-5c9gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.934348 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e681ef02-7b7d-40c7-9a6c-16d777364891" (UID: "e681ef02-7b7d-40c7-9a6c-16d777364891"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:22 crc kubenswrapper[4945]: I1008 15:36:22.985649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e681ef02-7b7d-40c7-9a6c-16d777364891" (UID: "e681ef02-7b7d-40c7-9a6c-16d777364891"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.003000 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.003034 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.003046 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.003054 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e681ef02-7b7d-40c7-9a6c-16d777364891-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.003063 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c9gc\" (UniqueName: \"kubernetes.io/projected/e681ef02-7b7d-40c7-9a6c-16d777364891-kube-api-access-5c9gc\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.003073 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.007788 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-config-data" (OuterVolumeSpecName: "config-data") pod "e681ef02-7b7d-40c7-9a6c-16d777364891" (UID: "e681ef02-7b7d-40c7-9a6c-16d777364891"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.104639 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e681ef02-7b7d-40c7-9a6c-16d777364891-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.136493 4945 generic.go:334] "Generic (PLEG): container finished" podID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerID="d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0" exitCode=0 Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.136537 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerDied","Data":"d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0"} Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.136557 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.136566 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e681ef02-7b7d-40c7-9a6c-16d777364891","Type":"ContainerDied","Data":"1ad05614a57d989096a76edf13456f46825be5be77f57a1aa051fc289b69d7e1"} Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.136583 4945 scope.go:117] "RemoveContainer" containerID="c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.160301 4945 scope.go:117] "RemoveContainer" containerID="ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.187251 4945 scope.go:117] "RemoveContainer" containerID="d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.187748 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.204965 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.229770 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.230348 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-api" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230365 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-api" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.230385 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-central-agent" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230393 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-central-agent" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.230412 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="proxy-httpd" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230421 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="proxy-httpd" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.230445 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-notification-agent" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230455 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-notification-agent" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.230479 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="sg-core" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230487 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="sg-core" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.230509 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-httpd" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230517 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-httpd" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230742 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-notification-agent" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230765 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-httpd" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230777 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="sg-core" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230790 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="48be9df6-e6b0-4345-8237-ea4a52cd03f7" containerName="neutron-api" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230809 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="ceilometer-central-agent" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.230819 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" containerName="proxy-httpd" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.233046 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.243267 4945 scope.go:117] "RemoveContainer" containerID="8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.244214 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.244357 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.244651 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.271519 4945 scope.go:117] "RemoveContainer" containerID="c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.271890 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27\": container with ID starting with c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27 not found: ID does not exist" containerID="c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.272023 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27"} err="failed to get container status \"c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27\": rpc error: code = NotFound desc = could not find container \"c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27\": container with ID starting with c23c8908eba226e86c65920e02d3026061c3522c12171426e28e9bb2a9d9cb27 not found: ID does not exist" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.272146 4945 scope.go:117] "RemoveContainer" containerID="ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.272494 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242\": container with ID starting with ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242 not found: ID does not exist" containerID="ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.272527 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242"} err="failed to get container status \"ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242\": rpc error: code = NotFound desc = could not find container \"ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242\": container with ID starting with ffdd70a9dab765cad85d2843fb5607387e56ba0e6b0a58de8a6b9054ee53b242 not found: ID does not exist" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.272546 4945 scope.go:117] "RemoveContainer" containerID="d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.272728 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0\": container with ID starting with d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0 not found: ID does not exist" containerID="d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.272750 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0"} err="failed to get container status \"d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0\": rpc error: code = NotFound desc = could not find container \"d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0\": container with ID starting with d4d8f25637405d74c1bf354859b9daed823455e011d92282133f2a811c2e3cf0 not found: ID does not exist" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.272763 4945 scope.go:117] "RemoveContainer" containerID="8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15" Oct 08 15:36:23 crc kubenswrapper[4945]: E1008 15:36:23.272943 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15\": container with ID starting with 8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15 not found: ID does not exist" containerID="8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.272961 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15"} err="failed to get container status \"8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15\": rpc error: code = NotFound desc = could not find container \"8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15\": container with ID starting with 8af963a0ba4fba2d765161cac8d963b371f8b0af3c1946dc2377cfea82542d15 not found: ID does not exist" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.308439 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.308520 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-run-httpd\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.308598 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-scripts\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.308677 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-config-data\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.308729 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-log-httpd\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.308766 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.308829 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg9qv\" (UniqueName: \"kubernetes.io/projected/c08d4f98-b49f-4c3c-a986-0077acbc467b-kube-api-access-gg9qv\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411266 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-run-httpd\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411387 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-scripts\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411469 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-config-data\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411542 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-log-httpd\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411581 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg9qv\" (UniqueName: \"kubernetes.io/projected/c08d4f98-b49f-4c3c-a986-0077acbc467b-kube-api-access-gg9qv\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411724 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.411985 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-run-httpd\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.412401 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-log-httpd\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.417668 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-config-data\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.418086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-scripts\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.420706 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.420928 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.443847 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg9qv\" (UniqueName: \"kubernetes.io/projected/c08d4f98-b49f-4c3c-a986-0077acbc467b-kube-api-access-gg9qv\") pod \"ceilometer-0\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " pod="openstack/ceilometer-0" Oct 08 15:36:23 crc kubenswrapper[4945]: I1008 15:36:23.562981 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:24 crc kubenswrapper[4945]: I1008 15:36:24.039306 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e681ef02-7b7d-40c7-9a6c-16d777364891" path="/var/lib/kubelet/pods/e681ef02-7b7d-40c7-9a6c-16d777364891/volumes" Oct 08 15:36:24 crc kubenswrapper[4945]: I1008 15:36:24.051652 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:24 crc kubenswrapper[4945]: W1008 15:36:24.057943 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc08d4f98_b49f_4c3c_a986_0077acbc467b.slice/crio-1ba5089e95b50345b06299ac098ab7b97e145d00425ab8919bcaace24729a125 WatchSource:0}: Error finding container 1ba5089e95b50345b06299ac098ab7b97e145d00425ab8919bcaace24729a125: Status 404 returned error can't find the container with id 1ba5089e95b50345b06299ac098ab7b97e145d00425ab8919bcaace24729a125 Oct 08 15:36:24 crc kubenswrapper[4945]: I1008 15:36:24.149073 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerStarted","Data":"1ba5089e95b50345b06299ac098ab7b97e145d00425ab8919bcaace24729a125"} Oct 08 15:36:25 crc kubenswrapper[4945]: I1008 15:36:25.024211 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:36:25 crc kubenswrapper[4945]: I1008 15:36:25.161235 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerStarted","Data":"aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d"} Oct 08 15:36:25 crc kubenswrapper[4945]: I1008 15:36:25.161282 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerStarted","Data":"569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4"} Oct 08 15:36:26 crc kubenswrapper[4945]: I1008 15:36:26.178003 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerStarted","Data":"61a9e3556abf6eba2040ec62743a6be0f8b9864d047c2a6d5017a6e1aede5f28"} Oct 08 15:36:26 crc kubenswrapper[4945]: I1008 15:36:26.181714 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerStarted","Data":"e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d"} Oct 08 15:36:27 crc kubenswrapper[4945]: I1008 15:36:27.197945 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerStarted","Data":"bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2"} Oct 08 15:36:27 crc kubenswrapper[4945]: I1008 15:36:27.198311 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:36:30 crc kubenswrapper[4945]: I1008 15:36:30.580515 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.286292507 podStartE2EDuration="7.580492762s" podCreationTimestamp="2025-10-08 15:36:23 +0000 UTC" firstStartedPulling="2025-10-08 15:36:24.060882255 +0000 UTC m=+1393.414797156" lastFinishedPulling="2025-10-08 15:36:26.35508251 +0000 UTC m=+1395.708997411" observedRunningTime="2025-10-08 15:36:27.230283657 +0000 UTC m=+1396.584198558" watchObservedRunningTime="2025-10-08 15:36:30.580492762 +0000 UTC m=+1399.934407683" Oct 08 15:36:30 crc kubenswrapper[4945]: I1008 15:36:30.592857 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:30 crc kubenswrapper[4945]: I1008 15:36:30.594760 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-central-agent" containerID="cri-o://569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4" gracePeriod=30 Oct 08 15:36:30 crc kubenswrapper[4945]: I1008 15:36:30.594850 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="proxy-httpd" containerID="cri-o://bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2" gracePeriod=30 Oct 08 15:36:30 crc kubenswrapper[4945]: I1008 15:36:30.594850 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="sg-core" containerID="cri-o://e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d" gracePeriod=30 Oct 08 15:36:30 crc kubenswrapper[4945]: I1008 15:36:30.594850 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-notification-agent" containerID="cri-o://aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d" gracePeriod=30 Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.250303 4945 generic.go:334] "Generic (PLEG): container finished" podID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerID="bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2" exitCode=0 Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.250599 4945 generic.go:334] "Generic (PLEG): container finished" podID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerID="e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d" exitCode=2 Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.250612 4945 generic.go:334] "Generic (PLEG): container finished" podID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerID="aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d" exitCode=0 Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.250636 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerDied","Data":"bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2"} Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.250664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerDied","Data":"e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d"} Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.250677 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerDied","Data":"aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d"} Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.922627 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.992412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-config-data\") pod \"c08d4f98-b49f-4c3c-a986-0077acbc467b\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.992451 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-log-httpd\") pod \"c08d4f98-b49f-4c3c-a986-0077acbc467b\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.992517 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-sg-core-conf-yaml\") pod \"c08d4f98-b49f-4c3c-a986-0077acbc467b\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.992574 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg9qv\" (UniqueName: \"kubernetes.io/projected/c08d4f98-b49f-4c3c-a986-0077acbc467b-kube-api-access-gg9qv\") pod \"c08d4f98-b49f-4c3c-a986-0077acbc467b\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.992646 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-scripts\") pod \"c08d4f98-b49f-4c3c-a986-0077acbc467b\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.992673 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-run-httpd\") pod \"c08d4f98-b49f-4c3c-a986-0077acbc467b\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.992721 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-combined-ca-bundle\") pod \"c08d4f98-b49f-4c3c-a986-0077acbc467b\" (UID: \"c08d4f98-b49f-4c3c-a986-0077acbc467b\") " Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.993631 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c08d4f98-b49f-4c3c-a986-0077acbc467b" (UID: "c08d4f98-b49f-4c3c-a986-0077acbc467b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.993739 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c08d4f98-b49f-4c3c-a986-0077acbc467b" (UID: "c08d4f98-b49f-4c3c-a986-0077acbc467b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.998858 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c08d4f98-b49f-4c3c-a986-0077acbc467b-kube-api-access-gg9qv" (OuterVolumeSpecName: "kube-api-access-gg9qv") pod "c08d4f98-b49f-4c3c-a986-0077acbc467b" (UID: "c08d4f98-b49f-4c3c-a986-0077acbc467b"). InnerVolumeSpecName "kube-api-access-gg9qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:36:31 crc kubenswrapper[4945]: I1008 15:36:31.998980 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-scripts" (OuterVolumeSpecName: "scripts") pod "c08d4f98-b49f-4c3c-a986-0077acbc467b" (UID: "c08d4f98-b49f-4c3c-a986-0077acbc467b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.024130 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c08d4f98-b49f-4c3c-a986-0077acbc467b" (UID: "c08d4f98-b49f-4c3c-a986-0077acbc467b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.078305 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c08d4f98-b49f-4c3c-a986-0077acbc467b" (UID: "c08d4f98-b49f-4c3c-a986-0077acbc467b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.094889 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.094919 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.094933 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg9qv\" (UniqueName: \"kubernetes.io/projected/c08d4f98-b49f-4c3c-a986-0077acbc467b-kube-api-access-gg9qv\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.094944 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.094956 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c08d4f98-b49f-4c3c-a986-0077acbc467b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.094966 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.105951 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-config-data" (OuterVolumeSpecName: "config-data") pod "c08d4f98-b49f-4c3c-a986-0077acbc467b" (UID: "c08d4f98-b49f-4c3c-a986-0077acbc467b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.193580 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.194668 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.195855 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08d4f98-b49f-4c3c-a986-0077acbc467b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.240252 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.266472 4945 generic.go:334] "Generic (PLEG): container finished" podID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerID="569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4" exitCode=0 Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.266559 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerDied","Data":"569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4"} Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.266619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c08d4f98-b49f-4c3c-a986-0077acbc467b","Type":"ContainerDied","Data":"1ba5089e95b50345b06299ac098ab7b97e145d00425ab8919bcaace24729a125"} Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.266642 4945 scope.go:117] "RemoveContainer" containerID="bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.266879 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.270964 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" event={"ID":"59d1ed3b-11c9-4deb-9947-605022b6592d","Type":"ContainerStarted","Data":"1bb696a9615feeb5c6ec70ad2cb044ebc1fa3b9e578521b98b35de44ac2be442"} Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.292955 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" podStartSLOduration=2.879873124 podStartE2EDuration="34.292934543s" podCreationTimestamp="2025-10-08 15:35:58 +0000 UTC" firstStartedPulling="2025-10-08 15:35:59.762712898 +0000 UTC m=+1369.116627789" lastFinishedPulling="2025-10-08 15:36:31.175774297 +0000 UTC m=+1400.529689208" observedRunningTime="2025-10-08 15:36:32.292068581 +0000 UTC m=+1401.645983482" watchObservedRunningTime="2025-10-08 15:36:32.292934543 +0000 UTC m=+1401.646849444" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.297671 4945 scope.go:117] "RemoveContainer" containerID="e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.313893 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.317136 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.325737 4945 scope.go:117] "RemoveContainer" containerID="aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.338386 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.355612 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.356047 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="proxy-httpd" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356063 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="proxy-httpd" Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.356089 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-notification-agent" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356097 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-notification-agent" Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.356123 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-central-agent" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356130 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-central-agent" Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.356142 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="sg-core" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356148 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="sg-core" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356340 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="proxy-httpd" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356366 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-central-agent" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356374 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="sg-core" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.356387 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" containerName="ceilometer-notification-agent" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.358232 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.362929 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.363264 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.371809 4945 scope.go:117] "RemoveContainer" containerID="569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.379767 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.384878 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.398926 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.398989 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.399007 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f49gz\" (UniqueName: \"kubernetes.io/projected/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-kube-api-access-f49gz\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.399065 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-run-httpd\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.399085 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-log-httpd\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.399139 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-config-data\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.399203 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-scripts\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.403850 4945 scope.go:117] "RemoveContainer" containerID="bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2" Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.404432 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2\": container with ID starting with bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2 not found: ID does not exist" containerID="bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.404460 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2"} err="failed to get container status \"bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2\": rpc error: code = NotFound desc = could not find container \"bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2\": container with ID starting with bc68314afddb82f0567bbd3086a4332a98483c7ba23d826da14adaad026abdf2 not found: ID does not exist" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.404480 4945 scope.go:117] "RemoveContainer" containerID="e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d" Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.404824 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d\": container with ID starting with e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d not found: ID does not exist" containerID="e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.404847 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d"} err="failed to get container status \"e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d\": rpc error: code = NotFound desc = could not find container \"e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d\": container with ID starting with e4650fc86ddeb667046f89c8262c94e9d72e6815eb5f1d3495b91d990b55de5d not found: ID does not exist" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.404862 4945 scope.go:117] "RemoveContainer" containerID="aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d" Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.405139 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d\": container with ID starting with aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d not found: ID does not exist" containerID="aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.405161 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d"} err="failed to get container status \"aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d\": rpc error: code = NotFound desc = could not find container \"aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d\": container with ID starting with aad393e12aecf925055cf29cb1d11e9d7b2c56bbf5364f153facd9f80cf8e74d not found: ID does not exist" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.405173 4945 scope.go:117] "RemoveContainer" containerID="569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4" Oct 08 15:36:32 crc kubenswrapper[4945]: E1008 15:36:32.405361 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4\": container with ID starting with 569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4 not found: ID does not exist" containerID="569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.405388 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4"} err="failed to get container status \"569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4\": rpc error: code = NotFound desc = could not find container \"569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4\": container with ID starting with 569c539d050bcfa23526dc4f4873e7124a7663c30dff1c7bdccbfbc56293b2b4 not found: ID does not exist" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.500752 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.500788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f49gz\" (UniqueName: \"kubernetes.io/projected/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-kube-api-access-f49gz\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.500841 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-run-httpd\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.500865 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-log-httpd\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.500886 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-config-data\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.500933 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-scripts\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.500994 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.501869 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-log-httpd\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.502640 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-run-httpd\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.504598 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.505269 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-scripts\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.505431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.506627 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-config-data\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.521003 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f49gz\" (UniqueName: \"kubernetes.io/projected/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-kube-api-access-f49gz\") pod \"ceilometer-0\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " pod="openstack/ceilometer-0" Oct 08 15:36:32 crc kubenswrapper[4945]: I1008 15:36:32.697251 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:36:33 crc kubenswrapper[4945]: W1008 15:36:33.207680 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f27b41a_ab54_4f70_ad8a_5fc4d2642042.slice/crio-59ce6d24e47466a0dc879ce8dbf1c9750de26610b93a4ceaf053f2550b23e843 WatchSource:0}: Error finding container 59ce6d24e47466a0dc879ce8dbf1c9750de26610b93a4ceaf053f2550b23e843: Status 404 returned error can't find the container with id 59ce6d24e47466a0dc879ce8dbf1c9750de26610b93a4ceaf053f2550b23e843 Oct 08 15:36:33 crc kubenswrapper[4945]: I1008 15:36:33.210126 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:36:33 crc kubenswrapper[4945]: I1008 15:36:33.280636 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerStarted","Data":"59ce6d24e47466a0dc879ce8dbf1c9750de26610b93a4ceaf053f2550b23e843"} Oct 08 15:36:34 crc kubenswrapper[4945]: I1008 15:36:34.047774 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c08d4f98-b49f-4c3c-a986-0077acbc467b" path="/var/lib/kubelet/pods/c08d4f98-b49f-4c3c-a986-0077acbc467b/volumes" Oct 08 15:36:34 crc kubenswrapper[4945]: I1008 15:36:34.295844 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerStarted","Data":"77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6"} Oct 08 15:36:34 crc kubenswrapper[4945]: I1008 15:36:34.295895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerStarted","Data":"b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe"} Oct 08 15:36:34 crc kubenswrapper[4945]: I1008 15:36:34.295929 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" containerID="cri-o://61a9e3556abf6eba2040ec62743a6be0f8b9864d047c2a6d5017a6e1aede5f28" gracePeriod=30 Oct 08 15:36:36 crc kubenswrapper[4945]: I1008 15:36:36.346711 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerStarted","Data":"27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7"} Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.357342 4945 generic.go:334] "Generic (PLEG): container finished" podID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerID="61a9e3556abf6eba2040ec62743a6be0f8b9864d047c2a6d5017a6e1aede5f28" exitCode=0 Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.357422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerDied","Data":"61a9e3556abf6eba2040ec62743a6be0f8b9864d047c2a6d5017a6e1aede5f28"} Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.357769 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c464851c-cb2a-4b05-aac2-b3ee6f7c9182","Type":"ContainerDied","Data":"1bed85dc545d31d674ad29548530b2ee65272749454136545d2ff05e79b3ee6f"} Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.357789 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bed85dc545d31d674ad29548530b2ee65272749454136545d2ff05e79b3ee6f" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.357807 4945 scope.go:117] "RemoveContainer" containerID="6c9cb0bc7a851e5bf1e07d0214dd3b688a9092552b925b6b8bb6be9a08744f20" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.366080 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.402028 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vdq2\" (UniqueName: \"kubernetes.io/projected/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-kube-api-access-9vdq2\") pod \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.402239 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-logs\") pod \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.402300 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-combined-ca-bundle\") pod \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.402398 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-custom-prometheus-ca\") pod \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.402430 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-config-data\") pod \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\" (UID: \"c464851c-cb2a-4b05-aac2-b3ee6f7c9182\") " Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.403085 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-logs" (OuterVolumeSpecName: "logs") pod "c464851c-cb2a-4b05-aac2-b3ee6f7c9182" (UID: "c464851c-cb2a-4b05-aac2-b3ee6f7c9182"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.403228 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.412267 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-kube-api-access-9vdq2" (OuterVolumeSpecName: "kube-api-access-9vdq2") pod "c464851c-cb2a-4b05-aac2-b3ee6f7c9182" (UID: "c464851c-cb2a-4b05-aac2-b3ee6f7c9182"). InnerVolumeSpecName "kube-api-access-9vdq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.454116 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "c464851c-cb2a-4b05-aac2-b3ee6f7c9182" (UID: "c464851c-cb2a-4b05-aac2-b3ee6f7c9182"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.461961 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c464851c-cb2a-4b05-aac2-b3ee6f7c9182" (UID: "c464851c-cb2a-4b05-aac2-b3ee6f7c9182"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.481553 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-config-data" (OuterVolumeSpecName: "config-data") pod "c464851c-cb2a-4b05-aac2-b3ee6f7c9182" (UID: "c464851c-cb2a-4b05-aac2-b3ee6f7c9182"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.505476 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.505517 4945 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.505530 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:37 crc kubenswrapper[4945]: I1008 15:36:37.505541 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vdq2\" (UniqueName: \"kubernetes.io/projected/c464851c-cb2a-4b05-aac2-b3ee6f7c9182-kube-api-access-9vdq2\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.366785 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.372047 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerStarted","Data":"1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51"} Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.372987 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.399994 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.413325 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.434964 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:36:38 crc kubenswrapper[4945]: E1008 15:36:38.435361 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.435376 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: E1008 15:36:38.435408 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.435414 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: E1008 15:36:38.435432 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.435437 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: E1008 15:36:38.435458 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.435465 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.435624 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.435637 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.435644 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.436307 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.442530 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.446729 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.428619056 podStartE2EDuration="6.44670864s" podCreationTimestamp="2025-10-08 15:36:32 +0000 UTC" firstStartedPulling="2025-10-08 15:36:33.20991592 +0000 UTC m=+1402.563830821" lastFinishedPulling="2025-10-08 15:36:37.228005504 +0000 UTC m=+1406.581920405" observedRunningTime="2025-10-08 15:36:38.42957784 +0000 UTC m=+1407.783492741" watchObservedRunningTime="2025-10-08 15:36:38.44670864 +0000 UTC m=+1407.800623541" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.466188 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.524051 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.524146 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxrvj\" (UniqueName: \"kubernetes.io/projected/ba695d96-8b81-4753-bd49-4686716797d7-kube-api-access-lxrvj\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.524223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.524264 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba695d96-8b81-4753-bd49-4686716797d7-logs\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.524296 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-config-data\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.626234 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.626827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxrvj\" (UniqueName: \"kubernetes.io/projected/ba695d96-8b81-4753-bd49-4686716797d7-kube-api-access-lxrvj\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.627296 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.627775 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba695d96-8b81-4753-bd49-4686716797d7-logs\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.627919 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-config-data\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.628275 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba695d96-8b81-4753-bd49-4686716797d7-logs\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.631499 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.632411 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-config-data\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.632642 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ba695d96-8b81-4753-bd49-4686716797d7-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.642889 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxrvj\" (UniqueName: \"kubernetes.io/projected/ba695d96-8b81-4753-bd49-4686716797d7-kube-api-access-lxrvj\") pod \"watcher-decision-engine-0\" (UID: \"ba695d96-8b81-4753-bd49-4686716797d7\") " pod="openstack/watcher-decision-engine-0" Oct 08 15:36:38 crc kubenswrapper[4945]: I1008 15:36:38.757367 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:39 crc kubenswrapper[4945]: I1008 15:36:39.212317 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 15:36:39 crc kubenswrapper[4945]: W1008 15:36:39.219508 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba695d96_8b81_4753_bd49_4686716797d7.slice/crio-a5673da1bdd57ac13cad2f6f27137c9960e22db34523981bb5c327f364779c8a WatchSource:0}: Error finding container a5673da1bdd57ac13cad2f6f27137c9960e22db34523981bb5c327f364779c8a: Status 404 returned error can't find the container with id a5673da1bdd57ac13cad2f6f27137c9960e22db34523981bb5c327f364779c8a Oct 08 15:36:39 crc kubenswrapper[4945]: I1008 15:36:39.389531 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"ba695d96-8b81-4753-bd49-4686716797d7","Type":"ContainerStarted","Data":"a5673da1bdd57ac13cad2f6f27137c9960e22db34523981bb5c327f364779c8a"} Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.055173 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" path="/var/lib/kubelet/pods/c464851c-cb2a-4b05-aac2-b3ee6f7c9182/volumes" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.430082 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"ba695d96-8b81-4753-bd49-4686716797d7","Type":"ContainerStarted","Data":"3b711a42f4046a776dcaa7f1310708208688ab11e65d45750d7ebf27e2da1dd2"} Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.460045 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.460021059 podStartE2EDuration="2.460021059s" podCreationTimestamp="2025-10-08 15:36:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:36:40.448029029 +0000 UTC m=+1409.801943930" watchObservedRunningTime="2025-10-08 15:36:40.460021059 +0000 UTC m=+1409.813935960" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.907612 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hzcm9"] Oct 08 15:36:40 crc kubenswrapper[4945]: E1008 15:36:40.908004 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.908019 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.908227 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.908240 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c464851c-cb2a-4b05-aac2-b3ee6f7c9182" containerName="watcher-decision-engine" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.909568 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.938658 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hzcm9"] Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.982426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-utilities\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.982498 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-catalog-content\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:40 crc kubenswrapper[4945]: I1008 15:36:40.982567 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwmbp\" (UniqueName: \"kubernetes.io/projected/92a2aa1b-ab92-494a-868d-2dd93bf95f31-kube-api-access-bwmbp\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.084301 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-utilities\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.084587 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-catalog-content\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.084642 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwmbp\" (UniqueName: \"kubernetes.io/projected/92a2aa1b-ab92-494a-868d-2dd93bf95f31-kube-api-access-bwmbp\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.084946 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-utilities\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.085570 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-catalog-content\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.104834 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwmbp\" (UniqueName: \"kubernetes.io/projected/92a2aa1b-ab92-494a-868d-2dd93bf95f31-kube-api-access-bwmbp\") pod \"redhat-operators-hzcm9\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.233381 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:41 crc kubenswrapper[4945]: I1008 15:36:41.718093 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hzcm9"] Oct 08 15:36:42 crc kubenswrapper[4945]: I1008 15:36:42.453682 4945 generic.go:334] "Generic (PLEG): container finished" podID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerID="57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e" exitCode=0 Oct 08 15:36:42 crc kubenswrapper[4945]: I1008 15:36:42.453731 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzcm9" event={"ID":"92a2aa1b-ab92-494a-868d-2dd93bf95f31","Type":"ContainerDied","Data":"57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e"} Oct 08 15:36:42 crc kubenswrapper[4945]: I1008 15:36:42.453953 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzcm9" event={"ID":"92a2aa1b-ab92-494a-868d-2dd93bf95f31","Type":"ContainerStarted","Data":"ed1fea76be8a1a2233f18b3bba263ce895295468f9b1448af64f1776a016ea0b"} Oct 08 15:36:43 crc kubenswrapper[4945]: I1008 15:36:43.468283 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzcm9" event={"ID":"92a2aa1b-ab92-494a-868d-2dd93bf95f31","Type":"ContainerStarted","Data":"af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18"} Oct 08 15:36:44 crc kubenswrapper[4945]: I1008 15:36:44.486920 4945 generic.go:334] "Generic (PLEG): container finished" podID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerID="af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18" exitCode=0 Oct 08 15:36:44 crc kubenswrapper[4945]: I1008 15:36:44.487018 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzcm9" event={"ID":"92a2aa1b-ab92-494a-868d-2dd93bf95f31","Type":"ContainerDied","Data":"af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18"} Oct 08 15:36:46 crc kubenswrapper[4945]: I1008 15:36:46.506637 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzcm9" event={"ID":"92a2aa1b-ab92-494a-868d-2dd93bf95f31","Type":"ContainerStarted","Data":"baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656"} Oct 08 15:36:46 crc kubenswrapper[4945]: I1008 15:36:46.529827 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hzcm9" podStartSLOduration=3.242273067 podStartE2EDuration="6.529806917s" podCreationTimestamp="2025-10-08 15:36:40 +0000 UTC" firstStartedPulling="2025-10-08 15:36:42.456133627 +0000 UTC m=+1411.810048528" lastFinishedPulling="2025-10-08 15:36:45.743667467 +0000 UTC m=+1415.097582378" observedRunningTime="2025-10-08 15:36:46.522290078 +0000 UTC m=+1415.876204979" watchObservedRunningTime="2025-10-08 15:36:46.529806917 +0000 UTC m=+1415.883721818" Oct 08 15:36:48 crc kubenswrapper[4945]: I1008 15:36:48.758547 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:48 crc kubenswrapper[4945]: I1008 15:36:48.793163 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.184939 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.185011 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.185057 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.185894 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5311a5a778713e57615c5afa0552244f2b5dba47b435a164e79654109f1ba68"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.185953 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://d5311a5a778713e57615c5afa0552244f2b5dba47b435a164e79654109f1ba68" gracePeriod=600 Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.540917 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="d5311a5a778713e57615c5afa0552244f2b5dba47b435a164e79654109f1ba68" exitCode=0 Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.540980 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"d5311a5a778713e57615c5afa0552244f2b5dba47b435a164e79654109f1ba68"} Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.541329 4945 scope.go:117] "RemoveContainer" containerID="d35c98890277031b010217d13608d3edf99fe1fc8eaff9d1b2452633ef6288de" Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.541543 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:49 crc kubenswrapper[4945]: I1008 15:36:49.581709 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 08 15:36:50 crc kubenswrapper[4945]: I1008 15:36:50.552561 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0"} Oct 08 15:36:51 crc kubenswrapper[4945]: I1008 15:36:51.234924 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:51 crc kubenswrapper[4945]: I1008 15:36:51.234994 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:51 crc kubenswrapper[4945]: I1008 15:36:51.317440 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:51 crc kubenswrapper[4945]: I1008 15:36:51.605203 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:51 crc kubenswrapper[4945]: I1008 15:36:51.655546 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hzcm9"] Oct 08 15:36:53 crc kubenswrapper[4945]: I1008 15:36:53.582221 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hzcm9" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="registry-server" containerID="cri-o://baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656" gracePeriod=2 Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.121456 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.134125 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwmbp\" (UniqueName: \"kubernetes.io/projected/92a2aa1b-ab92-494a-868d-2dd93bf95f31-kube-api-access-bwmbp\") pod \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.134188 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-utilities\") pod \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.134219 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-catalog-content\") pod \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\" (UID: \"92a2aa1b-ab92-494a-868d-2dd93bf95f31\") " Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.135070 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-utilities" (OuterVolumeSpecName: "utilities") pod "92a2aa1b-ab92-494a-868d-2dd93bf95f31" (UID: "92a2aa1b-ab92-494a-868d-2dd93bf95f31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.139950 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a2aa1b-ab92-494a-868d-2dd93bf95f31-kube-api-access-bwmbp" (OuterVolumeSpecName: "kube-api-access-bwmbp") pod "92a2aa1b-ab92-494a-868d-2dd93bf95f31" (UID: "92a2aa1b-ab92-494a-868d-2dd93bf95f31"). InnerVolumeSpecName "kube-api-access-bwmbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.219265 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92a2aa1b-ab92-494a-868d-2dd93bf95f31" (UID: "92a2aa1b-ab92-494a-868d-2dd93bf95f31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.236634 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwmbp\" (UniqueName: \"kubernetes.io/projected/92a2aa1b-ab92-494a-868d-2dd93bf95f31-kube-api-access-bwmbp\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.236691 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.236710 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a2aa1b-ab92-494a-868d-2dd93bf95f31-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.599974 4945 generic.go:334] "Generic (PLEG): container finished" podID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerID="baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656" exitCode=0 Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.600038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzcm9" event={"ID":"92a2aa1b-ab92-494a-868d-2dd93bf95f31","Type":"ContainerDied","Data":"baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656"} Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.600088 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzcm9" event={"ID":"92a2aa1b-ab92-494a-868d-2dd93bf95f31","Type":"ContainerDied","Data":"ed1fea76be8a1a2233f18b3bba263ce895295468f9b1448af64f1776a016ea0b"} Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.600182 4945 scope.go:117] "RemoveContainer" containerID="baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.600194 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzcm9" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.645972 4945 scope.go:117] "RemoveContainer" containerID="af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.656274 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hzcm9"] Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.670068 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hzcm9"] Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.688996 4945 scope.go:117] "RemoveContainer" containerID="57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.747999 4945 scope.go:117] "RemoveContainer" containerID="baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656" Oct 08 15:36:54 crc kubenswrapper[4945]: E1008 15:36:54.748421 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656\": container with ID starting with baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656 not found: ID does not exist" containerID="baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.748469 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656"} err="failed to get container status \"baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656\": rpc error: code = NotFound desc = could not find container \"baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656\": container with ID starting with baa4207dfd06a6c8ec370c6802795db547625452f5245a3e75548270a36e4656 not found: ID does not exist" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.748502 4945 scope.go:117] "RemoveContainer" containerID="af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18" Oct 08 15:36:54 crc kubenswrapper[4945]: E1008 15:36:54.748742 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18\": container with ID starting with af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18 not found: ID does not exist" containerID="af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.748762 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18"} err="failed to get container status \"af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18\": rpc error: code = NotFound desc = could not find container \"af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18\": container with ID starting with af1e2d735ae25e4524911f9f90ef4e5fb085b1f9b2b14b84ce5072149422da18 not found: ID does not exist" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.748779 4945 scope.go:117] "RemoveContainer" containerID="57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e" Oct 08 15:36:54 crc kubenswrapper[4945]: E1008 15:36:54.748972 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e\": container with ID starting with 57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e not found: ID does not exist" containerID="57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e" Oct 08 15:36:54 crc kubenswrapper[4945]: I1008 15:36:54.748994 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e"} err="failed to get container status \"57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e\": rpc error: code = NotFound desc = could not find container \"57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e\": container with ID starting with 57c08a3870e8b1577c977764adcb6f63c23f6eadffe5132ffe59380c4bbfce4e not found: ID does not exist" Oct 08 15:36:55 crc kubenswrapper[4945]: I1008 15:36:55.610520 4945 generic.go:334] "Generic (PLEG): container finished" podID="59d1ed3b-11c9-4deb-9947-605022b6592d" containerID="1bb696a9615feeb5c6ec70ad2cb044ebc1fa3b9e578521b98b35de44ac2be442" exitCode=0 Oct 08 15:36:55 crc kubenswrapper[4945]: I1008 15:36:55.610817 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" event={"ID":"59d1ed3b-11c9-4deb-9947-605022b6592d","Type":"ContainerDied","Data":"1bb696a9615feeb5c6ec70ad2cb044ebc1fa3b9e578521b98b35de44ac2be442"} Oct 08 15:36:56 crc kubenswrapper[4945]: I1008 15:36:56.039953 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" path="/var/lib/kubelet/pods/92a2aa1b-ab92-494a-868d-2dd93bf95f31/volumes" Oct 08 15:36:56 crc kubenswrapper[4945]: I1008 15:36:56.994731 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.111274 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-scripts\") pod \"59d1ed3b-11c9-4deb-9947-605022b6592d\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.111388 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-config-data\") pod \"59d1ed3b-11c9-4deb-9947-605022b6592d\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.111541 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-combined-ca-bundle\") pod \"59d1ed3b-11c9-4deb-9947-605022b6592d\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.111573 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9xkz\" (UniqueName: \"kubernetes.io/projected/59d1ed3b-11c9-4deb-9947-605022b6592d-kube-api-access-l9xkz\") pod \"59d1ed3b-11c9-4deb-9947-605022b6592d\" (UID: \"59d1ed3b-11c9-4deb-9947-605022b6592d\") " Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.118006 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-scripts" (OuterVolumeSpecName: "scripts") pod "59d1ed3b-11c9-4deb-9947-605022b6592d" (UID: "59d1ed3b-11c9-4deb-9947-605022b6592d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.118856 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d1ed3b-11c9-4deb-9947-605022b6592d-kube-api-access-l9xkz" (OuterVolumeSpecName: "kube-api-access-l9xkz") pod "59d1ed3b-11c9-4deb-9947-605022b6592d" (UID: "59d1ed3b-11c9-4deb-9947-605022b6592d"). InnerVolumeSpecName "kube-api-access-l9xkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.140936 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-config-data" (OuterVolumeSpecName: "config-data") pod "59d1ed3b-11c9-4deb-9947-605022b6592d" (UID: "59d1ed3b-11c9-4deb-9947-605022b6592d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.166310 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59d1ed3b-11c9-4deb-9947-605022b6592d" (UID: "59d1ed3b-11c9-4deb-9947-605022b6592d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.214799 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.214856 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9xkz\" (UniqueName: \"kubernetes.io/projected/59d1ed3b-11c9-4deb-9947-605022b6592d-kube-api-access-l9xkz\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.214878 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.214899 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d1ed3b-11c9-4deb-9947-605022b6592d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.634281 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" event={"ID":"59d1ed3b-11c9-4deb-9947-605022b6592d","Type":"ContainerDied","Data":"f363267d111614e9ed2f0a43c30e884b74e690bd370ceaeb5815ef2093665f61"} Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.634644 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f363267d111614e9ed2f0a43c30e884b74e690bd370ceaeb5815ef2093665f61" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.634338 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jh7mz" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.780400 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 15:36:57 crc kubenswrapper[4945]: E1008 15:36:57.781091 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="extract-utilities" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.781147 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="extract-utilities" Oct 08 15:36:57 crc kubenswrapper[4945]: E1008 15:36:57.781164 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="registry-server" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.781179 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="registry-server" Oct 08 15:36:57 crc kubenswrapper[4945]: E1008 15:36:57.781206 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="extract-content" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.781218 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="extract-content" Oct 08 15:36:57 crc kubenswrapper[4945]: E1008 15:36:57.781241 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d1ed3b-11c9-4deb-9947-605022b6592d" containerName="nova-cell0-conductor-db-sync" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.781252 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d1ed3b-11c9-4deb-9947-605022b6592d" containerName="nova-cell0-conductor-db-sync" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.781564 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a2aa1b-ab92-494a-868d-2dd93bf95f31" containerName="registry-server" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.781606 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d1ed3b-11c9-4deb-9947-605022b6592d" containerName="nova-cell0-conductor-db-sync" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.782563 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.784637 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.784872 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r5k2t" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.795126 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.825887 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.826303 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.826568 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddscq\" (UniqueName: \"kubernetes.io/projected/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-kube-api-access-ddscq\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.928396 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.928494 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddscq\" (UniqueName: \"kubernetes.io/projected/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-kube-api-access-ddscq\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.928544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.933749 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.934063 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:57 crc kubenswrapper[4945]: I1008 15:36:57.958545 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddscq\" (UniqueName: \"kubernetes.io/projected/d1cf9a55-81f7-45e6-97b6-b05f4e7f07be-kube-api-access-ddscq\") pod \"nova-cell0-conductor-0\" (UID: \"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be\") " pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:58 crc kubenswrapper[4945]: I1008 15:36:58.108211 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:58 crc kubenswrapper[4945]: I1008 15:36:58.564360 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 15:36:58 crc kubenswrapper[4945]: W1008 15:36:58.571632 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1cf9a55_81f7_45e6_97b6_b05f4e7f07be.slice/crio-7eb4a28057b8cf55e5c3213d5bc8c2413f703ecda1f1363ec71b7c17d5012b36 WatchSource:0}: Error finding container 7eb4a28057b8cf55e5c3213d5bc8c2413f703ecda1f1363ec71b7c17d5012b36: Status 404 returned error can't find the container with id 7eb4a28057b8cf55e5c3213d5bc8c2413f703ecda1f1363ec71b7c17d5012b36 Oct 08 15:36:58 crc kubenswrapper[4945]: I1008 15:36:58.643823 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be","Type":"ContainerStarted","Data":"7eb4a28057b8cf55e5c3213d5bc8c2413f703ecda1f1363ec71b7c17d5012b36"} Oct 08 15:36:59 crc kubenswrapper[4945]: I1008 15:36:59.658277 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d1cf9a55-81f7-45e6-97b6-b05f4e7f07be","Type":"ContainerStarted","Data":"3925d6011364dd0d7bdbeb5dfea8f4ac796d69c7918a591fc78391009dee1034"} Oct 08 15:36:59 crc kubenswrapper[4945]: I1008 15:36:59.660165 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 15:36:59 crc kubenswrapper[4945]: I1008 15:36:59.692437 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.692407693 podStartE2EDuration="2.692407693s" podCreationTimestamp="2025-10-08 15:36:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:36:59.681572508 +0000 UTC m=+1429.035487449" watchObservedRunningTime="2025-10-08 15:36:59.692407693 +0000 UTC m=+1429.046322624" Oct 08 15:37:02 crc kubenswrapper[4945]: I1008 15:37:02.707593 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.144181 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.612866 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-tv6pp"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.614649 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.616986 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.625645 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-tv6pp"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.627783 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.657484 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.657569 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-scripts\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.657593 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmvfs\" (UniqueName: \"kubernetes.io/projected/6b2392b9-5374-4ee2-919d-545683412f06-kube-api-access-xmvfs\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.657662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-config-data\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.761715 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-scripts\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.761763 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmvfs\" (UniqueName: \"kubernetes.io/projected/6b2392b9-5374-4ee2-919d-545683412f06-kube-api-access-xmvfs\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.761843 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-config-data\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.761904 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.769607 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.770052 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-scripts\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.771335 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.773896 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.774541 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.776995 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-config-data\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.787421 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmvfs\" (UniqueName: \"kubernetes.io/projected/6b2392b9-5374-4ee2-919d-545683412f06-kube-api-access-xmvfs\") pod \"nova-cell0-cell-mapping-tv6pp\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.804318 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.870480 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcbzb\" (UniqueName: \"kubernetes.io/projected/08c5f27f-07f8-48e3-906a-08047f6c17da-kube-api-access-kcbzb\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.870530 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.870620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.893455 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.895773 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.901926 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.903377 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.909142 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.915376 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.915667 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.925866 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.948238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.965409 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.967023 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.973242 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976323 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xvhn\" (UniqueName: \"kubernetes.io/projected/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-kube-api-access-5xvhn\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976414 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-config-data\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976431 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ptll\" (UniqueName: \"kubernetes.io/projected/7ffde086-3363-47e1-a9a9-61667781417a-kube-api-access-6ptll\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976486 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976532 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-config-data\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976555 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcbzb\" (UniqueName: \"kubernetes.io/projected/08c5f27f-07f8-48e3-906a-08047f6c17da-kube-api-access-kcbzb\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976576 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.976611 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-logs\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:03 crc kubenswrapper[4945]: I1008 15:37:03.990918 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.004177 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.013700 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcbzb\" (UniqueName: \"kubernetes.io/projected/08c5f27f-07f8-48e3-906a-08047f6c17da-kube-api-access-kcbzb\") pod \"nova-cell1-novncproxy-0\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.019184 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.087957 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-config-data\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.088271 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.088385 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ptll\" (UniqueName: \"kubernetes.io/projected/7ffde086-3363-47e1-a9a9-61667781417a-kube-api-access-6ptll\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.088531 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-config-data\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.088633 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbfdn\" (UniqueName: \"kubernetes.io/projected/690568fa-6f91-4653-b85a-49af210a08b2-kube-api-access-xbfdn\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.088738 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.088898 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-config-data\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.089047 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.089201 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-logs\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.089372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690568fa-6f91-4653-b85a-49af210a08b2-logs\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.089488 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xvhn\" (UniqueName: \"kubernetes.io/projected/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-kube-api-access-5xvhn\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.096279 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-logs\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.107028 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-config-data\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.120554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-config-data\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.131835 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.153944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ptll\" (UniqueName: \"kubernetes.io/projected/7ffde086-3363-47e1-a9a9-61667781417a-kube-api-access-6ptll\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.155659 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xvhn\" (UniqueName: \"kubernetes.io/projected/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-kube-api-access-5xvhn\") pod \"nova-api-0\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.184846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.185863 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84d796b5f9-vqkd7"] Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.188033 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.194314 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.194399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690568fa-6f91-4653-b85a-49af210a08b2-logs\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.194492 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-config-data\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.194521 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbfdn\" (UniqueName: \"kubernetes.io/projected/690568fa-6f91-4653-b85a-49af210a08b2-kube-api-access-xbfdn\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.195402 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690568fa-6f91-4653-b85a-49af210a08b2-logs\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.196098 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.228261 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84d796b5f9-vqkd7"] Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.230958 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.232676 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-config-data\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.242094 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.242844 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbfdn\" (UniqueName: \"kubernetes.io/projected/690568fa-6f91-4653-b85a-49af210a08b2-kube-api-access-xbfdn\") pod \"nova-metadata-0\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.249778 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.304964 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-config\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.308488 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.308746 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-svc\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.308934 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-swift-storage-0\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.309269 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-sb\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.309358 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh6kz\" (UniqueName: \"kubernetes.io/projected/b402b060-866f-40db-a058-74a48be28f64-kube-api-access-qh6kz\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.411148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-svc\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.411215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-swift-storage-0\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.411322 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-sb\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.411378 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh6kz\" (UniqueName: \"kubernetes.io/projected/b402b060-866f-40db-a058-74a48be28f64-kube-api-access-qh6kz\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.411457 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-config\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.411544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.419967 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.434682 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-svc\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.435281 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh6kz\" (UniqueName: \"kubernetes.io/projected/b402b060-866f-40db-a058-74a48be28f64-kube-api-access-qh6kz\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.435658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-sb\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.437312 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-swift-storage-0\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.439001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-config\") pod \"dnsmasq-dns-84d796b5f9-vqkd7\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.520176 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.551950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:04 crc kubenswrapper[4945]: W1008 15:37:04.786414 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b2392b9_5374_4ee2_919d_545683412f06.slice/crio-481828e63d6c66c80777daff58d9ee51fbc65f6c02b779edf723b966230cdc6b WatchSource:0}: Error finding container 481828e63d6c66c80777daff58d9ee51fbc65f6c02b779edf723b966230cdc6b: Status 404 returned error can't find the container with id 481828e63d6c66c80777daff58d9ee51fbc65f6c02b779edf723b966230cdc6b Oct 08 15:37:04 crc kubenswrapper[4945]: I1008 15:37:04.787992 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-tv6pp"] Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.013954 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zmh84"] Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.015382 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.018985 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.019276 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.028985 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zmh84"] Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.071449 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:05 crc kubenswrapper[4945]: W1008 15:37:05.073799 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ffde086_3363_47e1_a9a9_61667781417a.slice/crio-664a97bde87798a1f205a268b0907d1b946bea1ddde60adaad69cbb8b5fb3545 WatchSource:0}: Error finding container 664a97bde87798a1f205a268b0907d1b946bea1ddde60adaad69cbb8b5fb3545: Status 404 returned error can't find the container with id 664a97bde87798a1f205a268b0907d1b946bea1ddde60adaad69cbb8b5fb3545 Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.131418 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.132028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-scripts\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.132077 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-config-data\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.133242 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhjjm\" (UniqueName: \"kubernetes.io/projected/eb4085df-bd3d-4f38-b206-2cdadcf006bf-kube-api-access-jhjjm\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.207022 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.214919 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.235878 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhjjm\" (UniqueName: \"kubernetes.io/projected/eb4085df-bd3d-4f38-b206-2cdadcf006bf-kube-api-access-jhjjm\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.236043 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.236084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-scripts\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.236135 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-config-data\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.243307 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.244561 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-scripts\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.249623 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-config-data\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.252728 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhjjm\" (UniqueName: \"kubernetes.io/projected/eb4085df-bd3d-4f38-b206-2cdadcf006bf-kube-api-access-jhjjm\") pod \"nova-cell1-conductor-db-sync-zmh84\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: W1008 15:37:05.332712 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb402b060_866f_40db_a058_74a48be28f64.slice/crio-440a722f4e385942156a2ec19a7b0e1a1e47c4d473edbe1c9d33f2f3f7c97bc3 WatchSource:0}: Error finding container 440a722f4e385942156a2ec19a7b0e1a1e47c4d473edbe1c9d33f2f3f7c97bc3: Status 404 returned error can't find the container with id 440a722f4e385942156a2ec19a7b0e1a1e47c4d473edbe1c9d33f2f3f7c97bc3 Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.333368 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84d796b5f9-vqkd7"] Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.336249 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.350068 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.771993 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tv6pp" event={"ID":"6b2392b9-5374-4ee2-919d-545683412f06","Type":"ContainerStarted","Data":"824d3c69503a0ec1cb20fc535ad482935e1ec3ef7f0790d2c4ade097a7d9399f"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.772274 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tv6pp" event={"ID":"6b2392b9-5374-4ee2-919d-545683412f06","Type":"ContainerStarted","Data":"481828e63d6c66c80777daff58d9ee51fbc65f6c02b779edf723b966230cdc6b"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.774696 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ffde086-3363-47e1-a9a9-61667781417a","Type":"ContainerStarted","Data":"664a97bde87798a1f205a268b0907d1b946bea1ddde60adaad69cbb8b5fb3545"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.777543 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"08c5f27f-07f8-48e3-906a-08047f6c17da","Type":"ContainerStarted","Data":"0a6122be53ccbd2ab34752c0f4bf7de371359f1bc604993d6a37841e8ef4b8ef"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.778776 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b630c5e-fdf0-4aa3-8125-24ab26a105b4","Type":"ContainerStarted","Data":"3f26737c6960fdde90ff8130b845b74a24c3166e02d23ac295708cf8475c3ce3"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.780723 4945 generic.go:334] "Generic (PLEG): container finished" podID="b402b060-866f-40db-a058-74a48be28f64" containerID="3c092b22a02b2318946c47e8e94c357b173c1b21c2aaa6acd1ca8c1b54c31c47" exitCode=0 Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.780776 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" event={"ID":"b402b060-866f-40db-a058-74a48be28f64","Type":"ContainerDied","Data":"3c092b22a02b2318946c47e8e94c357b173c1b21c2aaa6acd1ca8c1b54c31c47"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.781375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" event={"ID":"b402b060-866f-40db-a058-74a48be28f64","Type":"ContainerStarted","Data":"440a722f4e385942156a2ec19a7b0e1a1e47c4d473edbe1c9d33f2f3f7c97bc3"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.782501 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"690568fa-6f91-4653-b85a-49af210a08b2","Type":"ContainerStarted","Data":"826b69fb5ef70914c633f72fb3665eedceaec4b177104e8aad6895bcc94d7d8e"} Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.814471 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-tv6pp" podStartSLOduration=2.814442037 podStartE2EDuration="2.814442037s" podCreationTimestamp="2025-10-08 15:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:05.800987128 +0000 UTC m=+1435.154902029" watchObservedRunningTime="2025-10-08 15:37:05.814442037 +0000 UTC m=+1435.168356968" Oct 08 15:37:05 crc kubenswrapper[4945]: I1008 15:37:05.834032 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zmh84"] Oct 08 15:37:06 crc kubenswrapper[4945]: I1008 15:37:06.796260 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" event={"ID":"b402b060-866f-40db-a058-74a48be28f64","Type":"ContainerStarted","Data":"ee3647dd523a3881a0a0add10110fa5a1b57bca8f3ea7e735eff6f6a98d93b03"} Oct 08 15:37:06 crc kubenswrapper[4945]: I1008 15:37:06.797929 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:06 crc kubenswrapper[4945]: I1008 15:37:06.800986 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zmh84" event={"ID":"eb4085df-bd3d-4f38-b206-2cdadcf006bf","Type":"ContainerStarted","Data":"18655ccb6da9e928ef848ad30691e324ddbc0eabf0582d0dbad5cc1195ac3aa3"} Oct 08 15:37:06 crc kubenswrapper[4945]: I1008 15:37:06.801134 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zmh84" event={"ID":"eb4085df-bd3d-4f38-b206-2cdadcf006bf","Type":"ContainerStarted","Data":"7596547a43806f913613538c1ab382596f3de7327883dcf3f05c4664f0338381"} Oct 08 15:37:06 crc kubenswrapper[4945]: I1008 15:37:06.832350 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" podStartSLOduration=2.832334484 podStartE2EDuration="2.832334484s" podCreationTimestamp="2025-10-08 15:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:06.815723278 +0000 UTC m=+1436.169638179" watchObservedRunningTime="2025-10-08 15:37:06.832334484 +0000 UTC m=+1436.186249375" Oct 08 15:37:06 crc kubenswrapper[4945]: I1008 15:37:06.836983 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zmh84" podStartSLOduration=2.836963297 podStartE2EDuration="2.836963297s" podCreationTimestamp="2025-10-08 15:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:06.831686308 +0000 UTC m=+1436.185601209" watchObservedRunningTime="2025-10-08 15:37:06.836963297 +0000 UTC m=+1436.190878198" Oct 08 15:37:07 crc kubenswrapper[4945]: I1008 15:37:07.624607 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:07 crc kubenswrapper[4945]: I1008 15:37:07.652762 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.835700 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"08c5f27f-07f8-48e3-906a-08047f6c17da","Type":"ContainerStarted","Data":"bd261f78913c8998d8fcce217c2a1bdd1aae8f390bf628ebeb308940199ab21b"} Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.836366 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="08c5f27f-07f8-48e3-906a-08047f6c17da" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bd261f78913c8998d8fcce217c2a1bdd1aae8f390bf628ebeb308940199ab21b" gracePeriod=30 Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.838962 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b630c5e-fdf0-4aa3-8125-24ab26a105b4","Type":"ContainerStarted","Data":"b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5"} Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.838986 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b630c5e-fdf0-4aa3-8125-24ab26a105b4","Type":"ContainerStarted","Data":"e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c"} Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.843641 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"690568fa-6f91-4653-b85a-49af210a08b2","Type":"ContainerStarted","Data":"961e4eb1c39067e5213730d13f07cd502244cc1bd5fc60a3a86a1101373702b8"} Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.843686 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"690568fa-6f91-4653-b85a-49af210a08b2","Type":"ContainerStarted","Data":"e0efea4db561962f85c3e4d6c2d15f94fc87f25576a5072cfd05d4b6759eb040"} Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.843783 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-log" containerID="cri-o://e0efea4db561962f85c3e4d6c2d15f94fc87f25576a5072cfd05d4b6759eb040" gracePeriod=30 Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.844032 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-metadata" containerID="cri-o://961e4eb1c39067e5213730d13f07cd502244cc1bd5fc60a3a86a1101373702b8" gracePeriod=30 Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.846774 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ffde086-3363-47e1-a9a9-61667781417a","Type":"ContainerStarted","Data":"c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8"} Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.859154 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.917660498 podStartE2EDuration="5.859135639s" podCreationTimestamp="2025-10-08 15:37:03 +0000 UTC" firstStartedPulling="2025-10-08 15:37:05.234403346 +0000 UTC m=+1434.588318237" lastFinishedPulling="2025-10-08 15:37:08.175878477 +0000 UTC m=+1437.529793378" observedRunningTime="2025-10-08 15:37:08.853930372 +0000 UTC m=+1438.207845273" watchObservedRunningTime="2025-10-08 15:37:08.859135639 +0000 UTC m=+1438.213050540" Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.871417 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.771445016 podStartE2EDuration="5.871396479s" podCreationTimestamp="2025-10-08 15:37:03 +0000 UTC" firstStartedPulling="2025-10-08 15:37:05.075609957 +0000 UTC m=+1434.429524848" lastFinishedPulling="2025-10-08 15:37:08.17556141 +0000 UTC m=+1437.529476311" observedRunningTime="2025-10-08 15:37:08.867469033 +0000 UTC m=+1438.221383934" watchObservedRunningTime="2025-10-08 15:37:08.871396479 +0000 UTC m=+1438.225311380" Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.886386 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.063630375 podStartE2EDuration="5.886368505s" podCreationTimestamp="2025-10-08 15:37:03 +0000 UTC" firstStartedPulling="2025-10-08 15:37:05.35450043 +0000 UTC m=+1434.708415331" lastFinishedPulling="2025-10-08 15:37:08.17723856 +0000 UTC m=+1437.531153461" observedRunningTime="2025-10-08 15:37:08.885577515 +0000 UTC m=+1438.239492416" watchObservedRunningTime="2025-10-08 15:37:08.886368505 +0000 UTC m=+1438.240283406" Oct 08 15:37:08 crc kubenswrapper[4945]: I1008 15:37:08.902757 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9154170820000003 podStartE2EDuration="5.902737284s" podCreationTimestamp="2025-10-08 15:37:03 +0000 UTC" firstStartedPulling="2025-10-08 15:37:05.217273257 +0000 UTC m=+1434.571188158" lastFinishedPulling="2025-10-08 15:37:08.204593459 +0000 UTC m=+1437.558508360" observedRunningTime="2025-10-08 15:37:08.897944378 +0000 UTC m=+1438.251859279" watchObservedRunningTime="2025-10-08 15:37:08.902737284 +0000 UTC m=+1438.256652185" Oct 08 15:37:09 crc kubenswrapper[4945]: I1008 15:37:09.197234 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:09 crc kubenswrapper[4945]: I1008 15:37:09.250352 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 15:37:09 crc kubenswrapper[4945]: I1008 15:37:09.520753 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 15:37:09 crc kubenswrapper[4945]: I1008 15:37:09.521000 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 15:37:09 crc kubenswrapper[4945]: I1008 15:37:09.858576 4945 generic.go:334] "Generic (PLEG): container finished" podID="690568fa-6f91-4653-b85a-49af210a08b2" containerID="e0efea4db561962f85c3e4d6c2d15f94fc87f25576a5072cfd05d4b6759eb040" exitCode=143 Oct 08 15:37:09 crc kubenswrapper[4945]: I1008 15:37:09.858692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"690568fa-6f91-4653-b85a-49af210a08b2","Type":"ContainerDied","Data":"e0efea4db561962f85c3e4d6c2d15f94fc87f25576a5072cfd05d4b6759eb040"} Oct 08 15:37:10 crc kubenswrapper[4945]: I1008 15:37:10.430001 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:37:10 crc kubenswrapper[4945]: I1008 15:37:10.430473 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="4971563b-eba1-4306-a525-e088a5b3f754" containerName="kube-state-metrics" containerID="cri-o://0f074551af9386ef7057a1d44db3e22a5af2d5afa54c054273ef5b22e9eb855c" gracePeriod=30 Oct 08 15:37:10 crc kubenswrapper[4945]: I1008 15:37:10.875385 4945 generic.go:334] "Generic (PLEG): container finished" podID="4971563b-eba1-4306-a525-e088a5b3f754" containerID="0f074551af9386ef7057a1d44db3e22a5af2d5afa54c054273ef5b22e9eb855c" exitCode=2 Oct 08 15:37:10 crc kubenswrapper[4945]: I1008 15:37:10.875758 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4971563b-eba1-4306-a525-e088a5b3f754","Type":"ContainerDied","Data":"0f074551af9386ef7057a1d44db3e22a5af2d5afa54c054273ef5b22e9eb855c"} Oct 08 15:37:10 crc kubenswrapper[4945]: I1008 15:37:10.954612 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.085016 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p6s9\" (UniqueName: \"kubernetes.io/projected/4971563b-eba1-4306-a525-e088a5b3f754-kube-api-access-8p6s9\") pod \"4971563b-eba1-4306-a525-e088a5b3f754\" (UID: \"4971563b-eba1-4306-a525-e088a5b3f754\") " Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.090817 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4971563b-eba1-4306-a525-e088a5b3f754-kube-api-access-8p6s9" (OuterVolumeSpecName: "kube-api-access-8p6s9") pod "4971563b-eba1-4306-a525-e088a5b3f754" (UID: "4971563b-eba1-4306-a525-e088a5b3f754"). InnerVolumeSpecName "kube-api-access-8p6s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.188957 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p6s9\" (UniqueName: \"kubernetes.io/projected/4971563b-eba1-4306-a525-e088a5b3f754-kube-api-access-8p6s9\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.920140 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4971563b-eba1-4306-a525-e088a5b3f754","Type":"ContainerDied","Data":"6d371ce1f996cbc66dd72295bb0cf27f9de4ad398f4ed0034405d94e37acc11b"} Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.921037 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.921355 4945 scope.go:117] "RemoveContainer" containerID="0f074551af9386ef7057a1d44db3e22a5af2d5afa54c054273ef5b22e9eb855c" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.975833 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.982346 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.993715 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:37:11 crc kubenswrapper[4945]: E1008 15:37:11.994203 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4971563b-eba1-4306-a525-e088a5b3f754" containerName="kube-state-metrics" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.994227 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4971563b-eba1-4306-a525-e088a5b3f754" containerName="kube-state-metrics" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.994471 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4971563b-eba1-4306-a525-e088a5b3f754" containerName="kube-state-metrics" Oct 08 15:37:11 crc kubenswrapper[4945]: I1008 15:37:11.995324 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.002577 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.002627 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.002579 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.043115 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4971563b-eba1-4306-a525-e088a5b3f754" path="/var/lib/kubelet/pods/4971563b-eba1-4306-a525-e088a5b3f754/volumes" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.115920 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz5gv\" (UniqueName: \"kubernetes.io/projected/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-api-access-kz5gv\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.116032 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.116458 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.116521 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.218972 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.219726 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.219870 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.220222 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz5gv\" (UniqueName: \"kubernetes.io/projected/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-api-access-kz5gv\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.223975 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.224974 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.226045 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.237893 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz5gv\" (UniqueName: \"kubernetes.io/projected/803b9d32-2d37-4fd2-96fd-b4001daf26ed-kube-api-access-kz5gv\") pod \"kube-state-metrics-0\" (UID: \"803b9d32-2d37-4fd2-96fd-b4001daf26ed\") " pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.320125 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.555810 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.556481 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-central-agent" containerID="cri-o://b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe" gracePeriod=30 Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.556548 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-notification-agent" containerID="cri-o://77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6" gracePeriod=30 Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.556548 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="proxy-httpd" containerID="cri-o://1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51" gracePeriod=30 Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.556523 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="sg-core" containerID="cri-o://27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7" gracePeriod=30 Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.782160 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 15:37:12 crc kubenswrapper[4945]: W1008 15:37:12.788069 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod803b9d32_2d37_4fd2_96fd_b4001daf26ed.slice/crio-d3da00d94145880a170ccf845c01ff74bd6eeb73d50acb037a1bdb69e457e796 WatchSource:0}: Error finding container d3da00d94145880a170ccf845c01ff74bd6eeb73d50acb037a1bdb69e457e796: Status 404 returned error can't find the container with id d3da00d94145880a170ccf845c01ff74bd6eeb73d50acb037a1bdb69e457e796 Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.934762 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"803b9d32-2d37-4fd2-96fd-b4001daf26ed","Type":"ContainerStarted","Data":"d3da00d94145880a170ccf845c01ff74bd6eeb73d50acb037a1bdb69e457e796"} Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.942090 4945 generic.go:334] "Generic (PLEG): container finished" podID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerID="1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51" exitCode=0 Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.942153 4945 generic.go:334] "Generic (PLEG): container finished" podID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerID="27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7" exitCode=2 Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.942297 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerDied","Data":"1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51"} Oct 08 15:37:12 crc kubenswrapper[4945]: I1008 15:37:12.942396 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerDied","Data":"27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7"} Oct 08 15:37:13 crc kubenswrapper[4945]: I1008 15:37:13.956734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerDied","Data":"b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe"} Oct 08 15:37:13 crc kubenswrapper[4945]: I1008 15:37:13.957175 4945 generic.go:334] "Generic (PLEG): container finished" podID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerID="b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe" exitCode=0 Oct 08 15:37:13 crc kubenswrapper[4945]: I1008 15:37:13.958983 4945 generic.go:334] "Generic (PLEG): container finished" podID="6b2392b9-5374-4ee2-919d-545683412f06" containerID="824d3c69503a0ec1cb20fc535ad482935e1ec3ef7f0790d2c4ade097a7d9399f" exitCode=0 Oct 08 15:37:13 crc kubenswrapper[4945]: I1008 15:37:13.959036 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tv6pp" event={"ID":"6b2392b9-5374-4ee2-919d-545683412f06","Type":"ContainerDied","Data":"824d3c69503a0ec1cb20fc535ad482935e1ec3ef7f0790d2c4ade097a7d9399f"} Oct 08 15:37:13 crc kubenswrapper[4945]: I1008 15:37:13.962100 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"803b9d32-2d37-4fd2-96fd-b4001daf26ed","Type":"ContainerStarted","Data":"00838ba7b25386a9e0cf221a2dff5926f2d0babb3c2c7afade110f1d0464af21"} Oct 08 15:37:13 crc kubenswrapper[4945]: I1008 15:37:13.962303 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 15:37:13 crc kubenswrapper[4945]: I1008 15:37:13.995144 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.505822299 podStartE2EDuration="2.995099132s" podCreationTimestamp="2025-10-08 15:37:11 +0000 UTC" firstStartedPulling="2025-10-08 15:37:12.790702499 +0000 UTC m=+1442.144617410" lastFinishedPulling="2025-10-08 15:37:13.279979342 +0000 UTC m=+1442.633894243" observedRunningTime="2025-10-08 15:37:13.992834407 +0000 UTC m=+1443.346749318" watchObservedRunningTime="2025-10-08 15:37:13.995099132 +0000 UTC m=+1443.349014043" Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.231806 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.232237 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.250571 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.283717 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.554347 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.669480 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b84564489-l4jsm"] Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.669984 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" podUID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerName="dnsmasq-dns" containerID="cri-o://59f37be1ffe2d24e9b8c37cf9889ecea154b870df48daa0e792ddd1dd5c01383" gracePeriod=10 Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.974180 4945 generic.go:334] "Generic (PLEG): container finished" podID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerID="59f37be1ffe2d24e9b8c37cf9889ecea154b870df48daa0e792ddd1dd5c01383" exitCode=0 Oct 08 15:37:14 crc kubenswrapper[4945]: I1008 15:37:14.974260 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" event={"ID":"c801f135-2df2-4cc1-9f62-dba3929cf17e","Type":"ContainerDied","Data":"59f37be1ffe2d24e9b8c37cf9889ecea154b870df48daa0e792ddd1dd5c01383"} Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.020245 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.232618 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.319237 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.319280 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.394722 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-swift-storage-0\") pod \"c801f135-2df2-4cc1-9f62-dba3929cf17e\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.394810 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxw6n\" (UniqueName: \"kubernetes.io/projected/c801f135-2df2-4cc1-9f62-dba3929cf17e-kube-api-access-gxw6n\") pod \"c801f135-2df2-4cc1-9f62-dba3929cf17e\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.394972 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-svc\") pod \"c801f135-2df2-4cc1-9f62-dba3929cf17e\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.395057 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-sb\") pod \"c801f135-2df2-4cc1-9f62-dba3929cf17e\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.395079 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-config\") pod \"c801f135-2df2-4cc1-9f62-dba3929cf17e\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.395158 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-nb\") pod \"c801f135-2df2-4cc1-9f62-dba3929cf17e\" (UID: \"c801f135-2df2-4cc1-9f62-dba3929cf17e\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.406237 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c801f135-2df2-4cc1-9f62-dba3929cf17e-kube-api-access-gxw6n" (OuterVolumeSpecName: "kube-api-access-gxw6n") pod "c801f135-2df2-4cc1-9f62-dba3929cf17e" (UID: "c801f135-2df2-4cc1-9f62-dba3929cf17e"). InnerVolumeSpecName "kube-api-access-gxw6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.460699 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.462777 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c801f135-2df2-4cc1-9f62-dba3929cf17e" (UID: "c801f135-2df2-4cc1-9f62-dba3929cf17e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.476327 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c801f135-2df2-4cc1-9f62-dba3929cf17e" (UID: "c801f135-2df2-4cc1-9f62-dba3929cf17e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.499362 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.499393 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.499406 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxw6n\" (UniqueName: \"kubernetes.io/projected/c801f135-2df2-4cc1-9f62-dba3929cf17e-kube-api-access-gxw6n\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.504155 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c801f135-2df2-4cc1-9f62-dba3929cf17e" (UID: "c801f135-2df2-4cc1-9f62-dba3929cf17e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.507413 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-config" (OuterVolumeSpecName: "config") pod "c801f135-2df2-4cc1-9f62-dba3929cf17e" (UID: "c801f135-2df2-4cc1-9f62-dba3929cf17e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.525542 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c801f135-2df2-4cc1-9f62-dba3929cf17e" (UID: "c801f135-2df2-4cc1-9f62-dba3929cf17e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.600645 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-config-data\") pod \"6b2392b9-5374-4ee2-919d-545683412f06\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.600775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmvfs\" (UniqueName: \"kubernetes.io/projected/6b2392b9-5374-4ee2-919d-545683412f06-kube-api-access-xmvfs\") pod \"6b2392b9-5374-4ee2-919d-545683412f06\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.600819 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-scripts\") pod \"6b2392b9-5374-4ee2-919d-545683412f06\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.600924 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-combined-ca-bundle\") pod \"6b2392b9-5374-4ee2-919d-545683412f06\" (UID: \"6b2392b9-5374-4ee2-919d-545683412f06\") " Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.601553 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.601581 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.601596 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c801f135-2df2-4cc1-9f62-dba3929cf17e-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.609948 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-scripts" (OuterVolumeSpecName: "scripts") pod "6b2392b9-5374-4ee2-919d-545683412f06" (UID: "6b2392b9-5374-4ee2-919d-545683412f06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.609975 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2392b9-5374-4ee2-919d-545683412f06-kube-api-access-xmvfs" (OuterVolumeSpecName: "kube-api-access-xmvfs") pod "6b2392b9-5374-4ee2-919d-545683412f06" (UID: "6b2392b9-5374-4ee2-919d-545683412f06"). InnerVolumeSpecName "kube-api-access-xmvfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.635363 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-config-data" (OuterVolumeSpecName: "config-data") pod "6b2392b9-5374-4ee2-919d-545683412f06" (UID: "6b2392b9-5374-4ee2-919d-545683412f06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.641023 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b2392b9-5374-4ee2-919d-545683412f06" (UID: "6b2392b9-5374-4ee2-919d-545683412f06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.703277 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.703309 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmvfs\" (UniqueName: \"kubernetes.io/projected/6b2392b9-5374-4ee2-919d-545683412f06-kube-api-access-xmvfs\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.703319 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.703327 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2392b9-5374-4ee2-919d-545683412f06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.965299 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.985977 4945 generic.go:334] "Generic (PLEG): container finished" podID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerID="77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6" exitCode=0 Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.986099 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerDied","Data":"77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6"} Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.986183 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f27b41a-ab54-4f70-ad8a-5fc4d2642042","Type":"ContainerDied","Data":"59ce6d24e47466a0dc879ce8dbf1c9750de26610b93a4ceaf053f2550b23e843"} Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.986231 4945 scope.go:117] "RemoveContainer" containerID="1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.986623 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.988252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tv6pp" event={"ID":"6b2392b9-5374-4ee2-919d-545683412f06","Type":"ContainerDied","Data":"481828e63d6c66c80777daff58d9ee51fbc65f6c02b779edf723b966230cdc6b"} Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.988265 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tv6pp" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.988277 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="481828e63d6c66c80777daff58d9ee51fbc65f6c02b779edf723b966230cdc6b" Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.992225 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" event={"ID":"c801f135-2df2-4cc1-9f62-dba3929cf17e","Type":"ContainerDied","Data":"534672ab02f381e7dbd1cd260ec7a5a53a521aa86707a8445a8bbec8defa83fc"} Oct 08 15:37:15 crc kubenswrapper[4945]: I1008 15:37:15.992350 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b84564489-l4jsm" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.028964 4945 scope.go:117] "RemoveContainer" containerID="27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.054810 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b84564489-l4jsm"] Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.062903 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b84564489-l4jsm"] Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.062927 4945 scope.go:117] "RemoveContainer" containerID="77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.083530 4945 scope.go:117] "RemoveContainer" containerID="b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.103938 4945 scope.go:117] "RemoveContainer" containerID="1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.104369 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51\": container with ID starting with 1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51 not found: ID does not exist" containerID="1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.104422 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51"} err="failed to get container status \"1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51\": rpc error: code = NotFound desc = could not find container \"1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51\": container with ID starting with 1c0383fadc2bf901dedeac53c6cee9b29fd8f0e9703b173adb96d89ad3cacd51 not found: ID does not exist" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.104456 4945 scope.go:117] "RemoveContainer" containerID="27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.104769 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7\": container with ID starting with 27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7 not found: ID does not exist" containerID="27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.104815 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7"} err="failed to get container status \"27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7\": rpc error: code = NotFound desc = could not find container \"27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7\": container with ID starting with 27fed3aeeb31d6a2616ad5fb008e51058077f0cad465739904af4fbfb1edbcf7 not found: ID does not exist" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.104842 4945 scope.go:117] "RemoveContainer" containerID="77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.105852 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6\": container with ID starting with 77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6 not found: ID does not exist" containerID="77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.105931 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6"} err="failed to get container status \"77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6\": rpc error: code = NotFound desc = could not find container \"77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6\": container with ID starting with 77fefe74eb23dfe300afcb71e075945243a3c0cb0d0bf2a3261a319f5c4449d6 not found: ID does not exist" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.105967 4945 scope.go:117] "RemoveContainer" containerID="b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.106257 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe\": container with ID starting with b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe not found: ID does not exist" containerID="b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.106291 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe"} err="failed to get container status \"b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe\": rpc error: code = NotFound desc = could not find container \"b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe\": container with ID starting with b7948748e1a81558bee39588992d9916c787220a1537960711f76eb9ee184afe not found: ID does not exist" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.106313 4945 scope.go:117] "RemoveContainer" containerID="59f37be1ffe2d24e9b8c37cf9889ecea154b870df48daa0e792ddd1dd5c01383" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.109947 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-combined-ca-bundle\") pod \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110463 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-run-httpd\") pod \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-log-httpd\") pod \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110563 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-sg-core-conf-yaml\") pod \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110642 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-config-data\") pod \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110687 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f49gz\" (UniqueName: \"kubernetes.io/projected/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-kube-api-access-f49gz\") pod \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110751 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-scripts\") pod \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\" (UID: \"4f27b41a-ab54-4f70-ad8a-5fc4d2642042\") " Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110911 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4f27b41a-ab54-4f70-ad8a-5fc4d2642042" (UID: "4f27b41a-ab54-4f70-ad8a-5fc4d2642042"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.110961 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4f27b41a-ab54-4f70-ad8a-5fc4d2642042" (UID: "4f27b41a-ab54-4f70-ad8a-5fc4d2642042"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.126637 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-kube-api-access-f49gz" (OuterVolumeSpecName: "kube-api-access-f49gz") pod "4f27b41a-ab54-4f70-ad8a-5fc4d2642042" (UID: "4f27b41a-ab54-4f70-ad8a-5fc4d2642042"). InnerVolumeSpecName "kube-api-access-f49gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.135443 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-scripts" (OuterVolumeSpecName: "scripts") pod "4f27b41a-ab54-4f70-ad8a-5fc4d2642042" (UID: "4f27b41a-ab54-4f70-ad8a-5fc4d2642042"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.135670 4945 scope.go:117] "RemoveContainer" containerID="1f429ec667ec642585a50f8a5121fe7c9e1361690f688476e8976d114b401757" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.198402 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4f27b41a-ab54-4f70-ad8a-5fc4d2642042" (UID: "4f27b41a-ab54-4f70-ad8a-5fc4d2642042"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.210699 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.213660 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.213685 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.213695 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.213707 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.213715 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f49gz\" (UniqueName: \"kubernetes.io/projected/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-kube-api-access-f49gz\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.218626 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.218832 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-log" containerID="cri-o://e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c" gracePeriod=30 Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.218958 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-api" containerID="cri-o://b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5" gracePeriod=30 Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.269969 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f27b41a-ab54-4f70-ad8a-5fc4d2642042" (UID: "4f27b41a-ab54-4f70-ad8a-5fc4d2642042"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.276372 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-config-data" (OuterVolumeSpecName: "config-data") pod "4f27b41a-ab54-4f70-ad8a-5fc4d2642042" (UID: "4f27b41a-ab54-4f70-ad8a-5fc4d2642042"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.315964 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.316262 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f27b41a-ab54-4f70-ad8a-5fc4d2642042-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.333908 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.347767 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.360473 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.360926 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerName="dnsmasq-dns" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.360938 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerName="dnsmasq-dns" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.360962 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerName="init" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.360969 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerName="init" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.360986 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="proxy-httpd" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.360992 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="proxy-httpd" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.361004 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-notification-agent" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361011 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-notification-agent" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.361022 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-central-agent" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361028 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-central-agent" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.361038 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="sg-core" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361043 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="sg-core" Oct 08 15:37:16 crc kubenswrapper[4945]: E1008 15:37:16.361055 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2392b9-5374-4ee2-919d-545683412f06" containerName="nova-manage" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361061 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2392b9-5374-4ee2-919d-545683412f06" containerName="nova-manage" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361265 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2392b9-5374-4ee2-919d-545683412f06" containerName="nova-manage" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361290 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="sg-core" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361304 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-notification-agent" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361315 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c801f135-2df2-4cc1-9f62-dba3929cf17e" containerName="dnsmasq-dns" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361323 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="ceilometer-central-agent" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.361335 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" containerName="proxy-httpd" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.363151 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.371341 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.371376 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.371426 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.372088 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.520576 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.520621 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.520705 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-scripts\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.520741 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx476\" (UniqueName: \"kubernetes.io/projected/0908392e-ee1c-4220-a893-0da9b12f813b-kube-api-access-tx476\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.520890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-run-httpd\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.521057 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-log-httpd\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.521094 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.521418 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-config-data\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.623767 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-config-data\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.623887 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.623911 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.624522 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-scripts\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.624551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx476\" (UniqueName: \"kubernetes.io/projected/0908392e-ee1c-4220-a893-0da9b12f813b-kube-api-access-tx476\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.624575 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-run-httpd\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.624646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-log-httpd\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.624665 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.625140 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-log-httpd\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.625268 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-run-httpd\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.627776 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-config-data\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.628948 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.629356 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.629429 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.630821 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-scripts\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.641922 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx476\" (UniqueName: \"kubernetes.io/projected/0908392e-ee1c-4220-a893-0da9b12f813b-kube-api-access-tx476\") pod \"ceilometer-0\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " pod="openstack/ceilometer-0" Oct 08 15:37:16 crc kubenswrapper[4945]: I1008 15:37:16.694457 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:17 crc kubenswrapper[4945]: I1008 15:37:17.004411 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerID="e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c" exitCode=143 Oct 08 15:37:17 crc kubenswrapper[4945]: I1008 15:37:17.004628 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b630c5e-fdf0-4aa3-8125-24ab26a105b4","Type":"ContainerDied","Data":"e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c"} Oct 08 15:37:17 crc kubenswrapper[4945]: I1008 15:37:17.009639 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7ffde086-3363-47e1-a9a9-61667781417a" containerName="nova-scheduler-scheduler" containerID="cri-o://c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8" gracePeriod=30 Oct 08 15:37:17 crc kubenswrapper[4945]: I1008 15:37:17.192883 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:18 crc kubenswrapper[4945]: I1008 15:37:18.023018 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerStarted","Data":"00a9b58478c062a2190516333e7ba8ce54b4580668f5d309fa459da9eb5f6743"} Oct 08 15:37:18 crc kubenswrapper[4945]: I1008 15:37:18.034948 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f27b41a-ab54-4f70-ad8a-5fc4d2642042" path="/var/lib/kubelet/pods/4f27b41a-ab54-4f70-ad8a-5fc4d2642042/volumes" Oct 08 15:37:18 crc kubenswrapper[4945]: I1008 15:37:18.035698 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c801f135-2df2-4cc1-9f62-dba3929cf17e" path="/var/lib/kubelet/pods/c801f135-2df2-4cc1-9f62-dba3929cf17e/volumes" Oct 08 15:37:18 crc kubenswrapper[4945]: I1008 15:37:18.036281 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerStarted","Data":"c9a8c80ddd1aa76aa4e1b46b53c12afafc8bfce467023778ceb205543f315391"} Oct 08 15:37:18 crc kubenswrapper[4945]: I1008 15:37:18.036309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerStarted","Data":"efa1d3aa7a6a3a224760ed859e2e059a032cf219e36ccabb435166097da04fdc"} Oct 08 15:37:19 crc kubenswrapper[4945]: I1008 15:37:19.032348 4945 generic.go:334] "Generic (PLEG): container finished" podID="eb4085df-bd3d-4f38-b206-2cdadcf006bf" containerID="18655ccb6da9e928ef848ad30691e324ddbc0eabf0582d0dbad5cc1195ac3aa3" exitCode=0 Oct 08 15:37:19 crc kubenswrapper[4945]: I1008 15:37:19.032456 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zmh84" event={"ID":"eb4085df-bd3d-4f38-b206-2cdadcf006bf","Type":"ContainerDied","Data":"18655ccb6da9e928ef848ad30691e324ddbc0eabf0582d0dbad5cc1195ac3aa3"} Oct 08 15:37:19 crc kubenswrapper[4945]: I1008 15:37:19.035851 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerStarted","Data":"a7111a26e0e8b428da8f322c818a89bb83840cdbf61aeb83d9a681262b4671bf"} Oct 08 15:37:19 crc kubenswrapper[4945]: E1008 15:37:19.252445 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 15:37:19 crc kubenswrapper[4945]: E1008 15:37:19.253839 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 15:37:19 crc kubenswrapper[4945]: E1008 15:37:19.255355 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 15:37:19 crc kubenswrapper[4945]: E1008 15:37:19.255387 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7ffde086-3363-47e1-a9a9-61667781417a" containerName="nova-scheduler-scheduler" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.454195 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.510863 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.615903 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhjjm\" (UniqueName: \"kubernetes.io/projected/eb4085df-bd3d-4f38-b206-2cdadcf006bf-kube-api-access-jhjjm\") pod \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.615990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-combined-ca-bundle\") pod \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.616015 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-scripts\") pod \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.616076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xvhn\" (UniqueName: \"kubernetes.io/projected/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-kube-api-access-5xvhn\") pod \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.616119 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-logs\") pod \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.616202 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-config-data\") pod \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.616235 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-combined-ca-bundle\") pod \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\" (UID: \"eb4085df-bd3d-4f38-b206-2cdadcf006bf\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.616287 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-config-data\") pod \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\" (UID: \"4b630c5e-fdf0-4aa3-8125-24ab26a105b4\") " Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.617335 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-logs" (OuterVolumeSpecName: "logs") pod "4b630c5e-fdf0-4aa3-8125-24ab26a105b4" (UID: "4b630c5e-fdf0-4aa3-8125-24ab26a105b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.624056 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-scripts" (OuterVolumeSpecName: "scripts") pod "eb4085df-bd3d-4f38-b206-2cdadcf006bf" (UID: "eb4085df-bd3d-4f38-b206-2cdadcf006bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.624716 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb4085df-bd3d-4f38-b206-2cdadcf006bf-kube-api-access-jhjjm" (OuterVolumeSpecName: "kube-api-access-jhjjm") pod "eb4085df-bd3d-4f38-b206-2cdadcf006bf" (UID: "eb4085df-bd3d-4f38-b206-2cdadcf006bf"). InnerVolumeSpecName "kube-api-access-jhjjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.624838 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-kube-api-access-5xvhn" (OuterVolumeSpecName: "kube-api-access-5xvhn") pod "4b630c5e-fdf0-4aa3-8125-24ab26a105b4" (UID: "4b630c5e-fdf0-4aa3-8125-24ab26a105b4"). InnerVolumeSpecName "kube-api-access-5xvhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.647768 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b630c5e-fdf0-4aa3-8125-24ab26a105b4" (UID: "4b630c5e-fdf0-4aa3-8125-24ab26a105b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.649553 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-config-data" (OuterVolumeSpecName: "config-data") pod "4b630c5e-fdf0-4aa3-8125-24ab26a105b4" (UID: "4b630c5e-fdf0-4aa3-8125-24ab26a105b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.653993 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb4085df-bd3d-4f38-b206-2cdadcf006bf" (UID: "eb4085df-bd3d-4f38-b206-2cdadcf006bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.659687 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-config-data" (OuterVolumeSpecName: "config-data") pod "eb4085df-bd3d-4f38-b206-2cdadcf006bf" (UID: "eb4085df-bd3d-4f38-b206-2cdadcf006bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718777 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhjjm\" (UniqueName: \"kubernetes.io/projected/eb4085df-bd3d-4f38-b206-2cdadcf006bf-kube-api-access-jhjjm\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718814 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718826 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718837 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xvhn\" (UniqueName: \"kubernetes.io/projected/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-kube-api-access-5xvhn\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718850 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718860 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718873 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4085df-bd3d-4f38-b206-2cdadcf006bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:20 crc kubenswrapper[4945]: I1008 15:37:20.718883 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b630c5e-fdf0-4aa3-8125-24ab26a105b4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.060289 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zmh84" event={"ID":"eb4085df-bd3d-4f38-b206-2cdadcf006bf","Type":"ContainerDied","Data":"7596547a43806f913613538c1ab382596f3de7327883dcf3f05c4664f0338381"} Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.060344 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7596547a43806f913613538c1ab382596f3de7327883dcf3f05c4664f0338381" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.060412 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zmh84" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.073860 4945 generic.go:334] "Generic (PLEG): container finished" podID="7ffde086-3363-47e1-a9a9-61667781417a" containerID="c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8" exitCode=0 Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.073938 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ffde086-3363-47e1-a9a9-61667781417a","Type":"ContainerDied","Data":"c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8"} Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.079966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerStarted","Data":"6852350f982566884c1e617f95195f371ca613a68c032dbbb963da9aae8b0dfb"} Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.081793 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.095157 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerID="b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5" exitCode=0 Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.095216 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b630c5e-fdf0-4aa3-8125-24ab26a105b4","Type":"ContainerDied","Data":"b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5"} Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.095232 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.095257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b630c5e-fdf0-4aa3-8125-24ab26a105b4","Type":"ContainerDied","Data":"3f26737c6960fdde90ff8130b845b74a24c3166e02d23ac295708cf8475c3ce3"} Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.095281 4945 scope.go:117] "RemoveContainer" containerID="b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.146570 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.444674637 podStartE2EDuration="5.146544645s" podCreationTimestamp="2025-10-08 15:37:16 +0000 UTC" firstStartedPulling="2025-10-08 15:37:17.192372483 +0000 UTC m=+1446.546287384" lastFinishedPulling="2025-10-08 15:37:19.894242491 +0000 UTC m=+1449.248157392" observedRunningTime="2025-10-08 15:37:21.107337917 +0000 UTC m=+1450.461252818" watchObservedRunningTime="2025-10-08 15:37:21.146544645 +0000 UTC m=+1450.500459546" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.207728 4945 scope.go:117] "RemoveContainer" containerID="e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.213893 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 15:37:21 crc kubenswrapper[4945]: E1008 15:37:21.214933 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb4085df-bd3d-4f38-b206-2cdadcf006bf" containerName="nova-cell1-conductor-db-sync" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.214956 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb4085df-bd3d-4f38-b206-2cdadcf006bf" containerName="nova-cell1-conductor-db-sync" Oct 08 15:37:21 crc kubenswrapper[4945]: E1008 15:37:21.214971 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-log" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.214977 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-log" Oct 08 15:37:21 crc kubenswrapper[4945]: E1008 15:37:21.214996 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-api" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.215001 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-api" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.221131 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb4085df-bd3d-4f38-b206-2cdadcf006bf" containerName="nova-cell1-conductor-db-sync" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.221177 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-log" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.221198 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" containerName="nova-api-api" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.221917 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.225557 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.231854 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.244274 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.253699 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.259838 4945 scope.go:117] "RemoveContainer" containerID="b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5" Oct 08 15:37:21 crc kubenswrapper[4945]: E1008 15:37:21.261377 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5\": container with ID starting with b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5 not found: ID does not exist" containerID="b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.261421 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5"} err="failed to get container status \"b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5\": rpc error: code = NotFound desc = could not find container \"b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5\": container with ID starting with b41869bdd733fb516041755e963d165a548357be9e2ba43da91bcf4f41a45da5 not found: ID does not exist" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.261447 4945 scope.go:117] "RemoveContainer" containerID="e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c" Oct 08 15:37:21 crc kubenswrapper[4945]: E1008 15:37:21.261824 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c\": container with ID starting with e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c not found: ID does not exist" containerID="e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.261860 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c"} err="failed to get container status \"e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c\": rpc error: code = NotFound desc = could not find container \"e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c\": container with ID starting with e65d4cf21f087764c0e759fcd9d78ef089c384e7780cff2a726954d715b8d79c not found: ID does not exist" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.264673 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.266466 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.269090 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.274357 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.358214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15d968-e006-4d49-92aa-97cabf627769-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.358369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15d968-e006-4d49-92aa-97cabf627769-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.358433 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqsnm\" (UniqueName: \"kubernetes.io/projected/ae15d968-e006-4d49-92aa-97cabf627769-kube-api-access-kqsnm\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.460340 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-config-data\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.460481 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l68tc\" (UniqueName: \"kubernetes.io/projected/26f7b93f-2750-47ef-a855-c3479a38759a-kube-api-access-l68tc\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.460568 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15d968-e006-4d49-92aa-97cabf627769-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.460629 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f7b93f-2750-47ef-a855-c3479a38759a-logs\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.460760 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15d968-e006-4d49-92aa-97cabf627769-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.460792 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.460874 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqsnm\" (UniqueName: \"kubernetes.io/projected/ae15d968-e006-4d49-92aa-97cabf627769-kube-api-access-kqsnm\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.471239 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15d968-e006-4d49-92aa-97cabf627769-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.473891 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15d968-e006-4d49-92aa-97cabf627769-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.482443 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqsnm\" (UniqueName: \"kubernetes.io/projected/ae15d968-e006-4d49-92aa-97cabf627769-kube-api-access-kqsnm\") pod \"nova-cell1-conductor-0\" (UID: \"ae15d968-e006-4d49-92aa-97cabf627769\") " pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.545655 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.576122 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l68tc\" (UniqueName: \"kubernetes.io/projected/26f7b93f-2750-47ef-a855-c3479a38759a-kube-api-access-l68tc\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.576230 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f7b93f-2750-47ef-a855-c3479a38759a-logs\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.576309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.576400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-config-data\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.577876 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f7b93f-2750-47ef-a855-c3479a38759a-logs\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.580621 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.583575 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-config-data\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.585586 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.600895 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l68tc\" (UniqueName: \"kubernetes.io/projected/26f7b93f-2750-47ef-a855-c3479a38759a-kube-api-access-l68tc\") pod \"nova-api-0\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.779889 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-config-data\") pod \"7ffde086-3363-47e1-a9a9-61667781417a\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.780368 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-combined-ca-bundle\") pod \"7ffde086-3363-47e1-a9a9-61667781417a\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.780565 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ptll\" (UniqueName: \"kubernetes.io/projected/7ffde086-3363-47e1-a9a9-61667781417a-kube-api-access-6ptll\") pod \"7ffde086-3363-47e1-a9a9-61667781417a\" (UID: \"7ffde086-3363-47e1-a9a9-61667781417a\") " Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.786340 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ffde086-3363-47e1-a9a9-61667781417a-kube-api-access-6ptll" (OuterVolumeSpecName: "kube-api-access-6ptll") pod "7ffde086-3363-47e1-a9a9-61667781417a" (UID: "7ffde086-3363-47e1-a9a9-61667781417a"). InnerVolumeSpecName "kube-api-access-6ptll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.815152 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ffde086-3363-47e1-a9a9-61667781417a" (UID: "7ffde086-3363-47e1-a9a9-61667781417a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.815273 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-config-data" (OuterVolumeSpecName: "config-data") pod "7ffde086-3363-47e1-a9a9-61667781417a" (UID: "7ffde086-3363-47e1-a9a9-61667781417a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.881078 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.883165 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ptll\" (UniqueName: \"kubernetes.io/projected/7ffde086-3363-47e1-a9a9-61667781417a-kube-api-access-6ptll\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.883203 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:21 crc kubenswrapper[4945]: I1008 15:37:21.883217 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffde086-3363-47e1-a9a9-61667781417a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.003820 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.043211 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b630c5e-fdf0-4aa3-8125-24ab26a105b4" path="/var/lib/kubelet/pods/4b630c5e-fdf0-4aa3-8125-24ab26a105b4/volumes" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.112607 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ffde086-3363-47e1-a9a9-61667781417a","Type":"ContainerDied","Data":"664a97bde87798a1f205a268b0907d1b946bea1ddde60adaad69cbb8b5fb3545"} Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.112760 4945 scope.go:117] "RemoveContainer" containerID="c94fd21f400acf607210ab4ccb4088b43ff00917c1b8da4ba2e48b39aa58c2c8" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.112920 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.123541 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae15d968-e006-4d49-92aa-97cabf627769","Type":"ContainerStarted","Data":"737953ea9eba19d16cb0bfb60d4c01525caf102e036c91c0a0b376e6fb7245b5"} Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.168466 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.189683 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.198968 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:22 crc kubenswrapper[4945]: E1008 15:37:22.199964 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ffde086-3363-47e1-a9a9-61667781417a" containerName="nova-scheduler-scheduler" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.199986 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ffde086-3363-47e1-a9a9-61667781417a" containerName="nova-scheduler-scheduler" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.201052 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ffde086-3363-47e1-a9a9-61667781417a" containerName="nova-scheduler-scheduler" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.203781 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.207149 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.214862 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.291093 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8pkp\" (UniqueName: \"kubernetes.io/projected/b3a9136d-3567-4523-97bb-56936b7f6ae4-kube-api-access-g8pkp\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.291143 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-config-data\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.291180 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.309891 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:22 crc kubenswrapper[4945]: W1008 15:37:22.315787 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26f7b93f_2750_47ef_a855_c3479a38759a.slice/crio-8e93f66f98f9155e3b2d7fb144bfc9de0ca354da679fa26df349db5324a63501 WatchSource:0}: Error finding container 8e93f66f98f9155e3b2d7fb144bfc9de0ca354da679fa26df349db5324a63501: Status 404 returned error can't find the container with id 8e93f66f98f9155e3b2d7fb144bfc9de0ca354da679fa26df349db5324a63501 Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.331555 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.394161 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8pkp\" (UniqueName: \"kubernetes.io/projected/b3a9136d-3567-4523-97bb-56936b7f6ae4-kube-api-access-g8pkp\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.394237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-config-data\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.394315 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.398797 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.399227 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-config-data\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.409991 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8pkp\" (UniqueName: \"kubernetes.io/projected/b3a9136d-3567-4523-97bb-56936b7f6ae4-kube-api-access-g8pkp\") pod \"nova-scheduler-0\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " pod="openstack/nova-scheduler-0" Oct 08 15:37:22 crc kubenswrapper[4945]: I1008 15:37:22.543472 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.015721 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:37:23 crc kubenswrapper[4945]: W1008 15:37:23.016431 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3a9136d_3567_4523_97bb_56936b7f6ae4.slice/crio-4cd71deaaa5b3d2bae801c7a8d9a17aa1ed7383230779fb26ff09cdd40ee9799 WatchSource:0}: Error finding container 4cd71deaaa5b3d2bae801c7a8d9a17aa1ed7383230779fb26ff09cdd40ee9799: Status 404 returned error can't find the container with id 4cd71deaaa5b3d2bae801c7a8d9a17aa1ed7383230779fb26ff09cdd40ee9799 Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.147868 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f7b93f-2750-47ef-a855-c3479a38759a","Type":"ContainerStarted","Data":"2cf7a2a1b470261f3b61e2c6ea3e6c1fedca423033cd911e3b3b8bb6e39eeddc"} Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.148072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f7b93f-2750-47ef-a855-c3479a38759a","Type":"ContainerStarted","Data":"be33f8dcd7a09c8c2fbf288b6b0a761ef7ac831c2f2ec6007c8d23a136d9b6cd"} Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.148178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f7b93f-2750-47ef-a855-c3479a38759a","Type":"ContainerStarted","Data":"8e93f66f98f9155e3b2d7fb144bfc9de0ca354da679fa26df349db5324a63501"} Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.162695 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3a9136d-3567-4523-97bb-56936b7f6ae4","Type":"ContainerStarted","Data":"4cd71deaaa5b3d2bae801c7a8d9a17aa1ed7383230779fb26ff09cdd40ee9799"} Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.166399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae15d968-e006-4d49-92aa-97cabf627769","Type":"ContainerStarted","Data":"58d104805d48ae1e9f7978610e6fa76a2ac86dc61272656dc3cfbe5a5cbbbb71"} Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.167501 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.168737 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.168722907 podStartE2EDuration="2.168722907s" podCreationTimestamp="2025-10-08 15:37:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:23.163231773 +0000 UTC m=+1452.517146694" watchObservedRunningTime="2025-10-08 15:37:23.168722907 +0000 UTC m=+1452.522637808" Oct 08 15:37:23 crc kubenswrapper[4945]: I1008 15:37:23.187221 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.187203618 podStartE2EDuration="2.187203618s" podCreationTimestamp="2025-10-08 15:37:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:23.181387896 +0000 UTC m=+1452.535302797" watchObservedRunningTime="2025-10-08 15:37:23.187203618 +0000 UTC m=+1452.541118519" Oct 08 15:37:24 crc kubenswrapper[4945]: I1008 15:37:24.036545 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ffde086-3363-47e1-a9a9-61667781417a" path="/var/lib/kubelet/pods/7ffde086-3363-47e1-a9a9-61667781417a/volumes" Oct 08 15:37:24 crc kubenswrapper[4945]: I1008 15:37:24.176871 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3a9136d-3567-4523-97bb-56936b7f6ae4","Type":"ContainerStarted","Data":"b9446e663432cd61b397f3728603ba4c2c67baddfdc63924a70e6024bf3f3e21"} Oct 08 15:37:24 crc kubenswrapper[4945]: I1008 15:37:24.197909 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.19789323 podStartE2EDuration="2.19789323s" podCreationTimestamp="2025-10-08 15:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:24.19133963 +0000 UTC m=+1453.545254541" watchObservedRunningTime="2025-10-08 15:37:24.19789323 +0000 UTC m=+1453.551808131" Oct 08 15:37:27 crc kubenswrapper[4945]: I1008 15:37:27.544365 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 15:37:31 crc kubenswrapper[4945]: I1008 15:37:31.591508 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 15:37:31 crc kubenswrapper[4945]: I1008 15:37:31.881440 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:37:31 crc kubenswrapper[4945]: I1008 15:37:31.881513 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:37:32 crc kubenswrapper[4945]: I1008 15:37:32.543808 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 15:37:32 crc kubenswrapper[4945]: I1008 15:37:32.586804 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 15:37:32 crc kubenswrapper[4945]: I1008 15:37:32.963340 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 15:37:32 crc kubenswrapper[4945]: I1008 15:37:32.963367 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 15:37:33 crc kubenswrapper[4945]: I1008 15:37:33.297713 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.356003 4945 generic.go:334] "Generic (PLEG): container finished" podID="08c5f27f-07f8-48e3-906a-08047f6c17da" containerID="bd261f78913c8998d8fcce217c2a1bdd1aae8f390bf628ebeb308940199ab21b" exitCode=137 Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.356088 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"08c5f27f-07f8-48e3-906a-08047f6c17da","Type":"ContainerDied","Data":"bd261f78913c8998d8fcce217c2a1bdd1aae8f390bf628ebeb308940199ab21b"} Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.356542 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"08c5f27f-07f8-48e3-906a-08047f6c17da","Type":"ContainerDied","Data":"0a6122be53ccbd2ab34752c0f4bf7de371359f1bc604993d6a37841e8ef4b8ef"} Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.357123 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a6122be53ccbd2ab34752c0f4bf7de371359f1bc604993d6a37841e8ef4b8ef" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.359675 4945 generic.go:334] "Generic (PLEG): container finished" podID="690568fa-6f91-4653-b85a-49af210a08b2" containerID="961e4eb1c39067e5213730d13f07cd502244cc1bd5fc60a3a86a1101373702b8" exitCode=137 Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.359729 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"690568fa-6f91-4653-b85a-49af210a08b2","Type":"ContainerDied","Data":"961e4eb1c39067e5213730d13f07cd502244cc1bd5fc60a3a86a1101373702b8"} Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.379701 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.504158 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.532038 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcbzb\" (UniqueName: \"kubernetes.io/projected/08c5f27f-07f8-48e3-906a-08047f6c17da-kube-api-access-kcbzb\") pod \"08c5f27f-07f8-48e3-906a-08047f6c17da\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.532153 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-config-data\") pod \"08c5f27f-07f8-48e3-906a-08047f6c17da\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.532321 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-combined-ca-bundle\") pod \"08c5f27f-07f8-48e3-906a-08047f6c17da\" (UID: \"08c5f27f-07f8-48e3-906a-08047f6c17da\") " Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.542697 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c5f27f-07f8-48e3-906a-08047f6c17da-kube-api-access-kcbzb" (OuterVolumeSpecName: "kube-api-access-kcbzb") pod "08c5f27f-07f8-48e3-906a-08047f6c17da" (UID: "08c5f27f-07f8-48e3-906a-08047f6c17da"). InnerVolumeSpecName "kube-api-access-kcbzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.566498 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08c5f27f-07f8-48e3-906a-08047f6c17da" (UID: "08c5f27f-07f8-48e3-906a-08047f6c17da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.568419 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-config-data" (OuterVolumeSpecName: "config-data") pod "08c5f27f-07f8-48e3-906a-08047f6c17da" (UID: "08c5f27f-07f8-48e3-906a-08047f6c17da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.635362 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-config-data\") pod \"690568fa-6f91-4653-b85a-49af210a08b2\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.635666 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbfdn\" (UniqueName: \"kubernetes.io/projected/690568fa-6f91-4653-b85a-49af210a08b2-kube-api-access-xbfdn\") pod \"690568fa-6f91-4653-b85a-49af210a08b2\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.635782 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-combined-ca-bundle\") pod \"690568fa-6f91-4653-b85a-49af210a08b2\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.635834 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690568fa-6f91-4653-b85a-49af210a08b2-logs\") pod \"690568fa-6f91-4653-b85a-49af210a08b2\" (UID: \"690568fa-6f91-4653-b85a-49af210a08b2\") " Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.636450 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690568fa-6f91-4653-b85a-49af210a08b2-logs" (OuterVolumeSpecName: "logs") pod "690568fa-6f91-4653-b85a-49af210a08b2" (UID: "690568fa-6f91-4653-b85a-49af210a08b2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.636988 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcbzb\" (UniqueName: \"kubernetes.io/projected/08c5f27f-07f8-48e3-906a-08047f6c17da-kube-api-access-kcbzb\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.637016 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.637033 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690568fa-6f91-4653-b85a-49af210a08b2-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.637052 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c5f27f-07f8-48e3-906a-08047f6c17da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.640494 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690568fa-6f91-4653-b85a-49af210a08b2-kube-api-access-xbfdn" (OuterVolumeSpecName: "kube-api-access-xbfdn") pod "690568fa-6f91-4653-b85a-49af210a08b2" (UID: "690568fa-6f91-4653-b85a-49af210a08b2"). InnerVolumeSpecName "kube-api-access-xbfdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.662283 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "690568fa-6f91-4653-b85a-49af210a08b2" (UID: "690568fa-6f91-4653-b85a-49af210a08b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.665452 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-config-data" (OuterVolumeSpecName: "config-data") pod "690568fa-6f91-4653-b85a-49af210a08b2" (UID: "690568fa-6f91-4653-b85a-49af210a08b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.739292 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.739345 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbfdn\" (UniqueName: \"kubernetes.io/projected/690568fa-6f91-4653-b85a-49af210a08b2-kube-api-access-xbfdn\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:39.739362 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690568fa-6f91-4653-b85a-49af210a08b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.375557 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.375575 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"690568fa-6f91-4653-b85a-49af210a08b2","Type":"ContainerDied","Data":"826b69fb5ef70914c633f72fb3665eedceaec4b177104e8aad6895bcc94d7d8e"} Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.375586 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.376199 4945 scope.go:117] "RemoveContainer" containerID="961e4eb1c39067e5213730d13f07cd502244cc1bd5fc60a3a86a1101373702b8" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.400846 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.410021 4945 scope.go:117] "RemoveContainer" containerID="e0efea4db561962f85c3e4d6c2d15f94fc87f25576a5072cfd05d4b6759eb040" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.411942 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.424092 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.440925 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.452013 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: E1008 15:37:40.452550 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-metadata" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.452566 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-metadata" Oct 08 15:37:40 crc kubenswrapper[4945]: E1008 15:37:40.452605 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c5f27f-07f8-48e3-906a-08047f6c17da" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.452614 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c5f27f-07f8-48e3-906a-08047f6c17da" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 15:37:40 crc kubenswrapper[4945]: E1008 15:37:40.452635 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-log" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.452642 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-log" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.452968 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-metadata" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.452996 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c5f27f-07f8-48e3-906a-08047f6c17da" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.453013 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="690568fa-6f91-4653-b85a-49af210a08b2" containerName="nova-metadata-log" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.453876 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.458726 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.459058 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.465827 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.481978 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.487875 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.491656 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.491841 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.501458 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.514201 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.655900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.655978 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656014 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-config-data\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656045 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656067 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a5605-d46e-4de8-bf48-0d1b32fa3971-logs\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656089 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656123 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656144 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg9d6\" (UniqueName: \"kubernetes.io/projected/505a5605-d46e-4de8-bf48-0d1b32fa3971-kube-api-access-cg9d6\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmbz5\" (UniqueName: \"kubernetes.io/projected/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-kube-api-access-gmbz5\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.656644 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.758952 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmbz5\" (UniqueName: \"kubernetes.io/projected/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-kube-api-access-gmbz5\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759046 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759091 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-config-data\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759189 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759206 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a5605-d46e-4de8-bf48-0d1b32fa3971-logs\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759229 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759246 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.759267 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg9d6\" (UniqueName: \"kubernetes.io/projected/505a5605-d46e-4de8-bf48-0d1b32fa3971-kube-api-access-cg9d6\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.760285 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a5605-d46e-4de8-bf48-0d1b32fa3971-logs\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.764350 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.764667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.765679 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.766968 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.767083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-config-data\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.767619 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.767643 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.782890 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg9d6\" (UniqueName: \"kubernetes.io/projected/505a5605-d46e-4de8-bf48-0d1b32fa3971-kube-api-access-cg9d6\") pod \"nova-metadata-0\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " pod="openstack/nova-metadata-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.789429 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmbz5\" (UniqueName: \"kubernetes.io/projected/7a3edb5a-ae0f-411f-b1f5-a9d591a831b6-kube-api-access-gmbz5\") pod \"nova-cell1-novncproxy-0\" (UID: \"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:40 crc kubenswrapper[4945]: I1008 15:37:40.808328 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.086189 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.250596 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:37:41 crc kubenswrapper[4945]: W1008 15:37:41.259919 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod505a5605_d46e_4de8_bf48_0d1b32fa3971.slice/crio-a868acf6cec282bbfb6b471bd12d132fa6d1c5a2bca743e8e8cd16b680011d3b WatchSource:0}: Error finding container a868acf6cec282bbfb6b471bd12d132fa6d1c5a2bca743e8e8cd16b680011d3b: Status 404 returned error can't find the container with id a868acf6cec282bbfb6b471bd12d132fa6d1c5a2bca743e8e8cd16b680011d3b Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.389585 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"505a5605-d46e-4de8-bf48-0d1b32fa3971","Type":"ContainerStarted","Data":"a868acf6cec282bbfb6b471bd12d132fa6d1c5a2bca743e8e8cd16b680011d3b"} Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.579938 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.887728 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.888516 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.898251 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 15:37:41 crc kubenswrapper[4945]: I1008 15:37:41.903204 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.035039 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08c5f27f-07f8-48e3-906a-08047f6c17da" path="/var/lib/kubelet/pods/08c5f27f-07f8-48e3-906a-08047f6c17da/volumes" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.035590 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690568fa-6f91-4653-b85a-49af210a08b2" path="/var/lib/kubelet/pods/690568fa-6f91-4653-b85a-49af210a08b2/volumes" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.405706 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"505a5605-d46e-4de8-bf48-0d1b32fa3971","Type":"ContainerStarted","Data":"1fbb94a8d8f61a0b6cef4a5fdb687b9157cf516718378738b1a4703dec50900b"} Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.405760 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"505a5605-d46e-4de8-bf48-0d1b32fa3971","Type":"ContainerStarted","Data":"4711b6a17f7f102d9385a94f060dcdb2664bb77fd86d1db33c0cec5823c0ee92"} Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.408710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6","Type":"ContainerStarted","Data":"4cb4446654fe062991b3186893ef2e28e823d4d1d52b46befe86dea4c1e4fc8b"} Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.409077 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7a3edb5a-ae0f-411f-b1f5-a9d591a831b6","Type":"ContainerStarted","Data":"44bb4dcd4c8c47cf3014074806bd74acca5fe15c8922085db1bb41d32c3c37dc"} Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.409491 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.421544 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.441782 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.441760733 podStartE2EDuration="2.441760733s" podCreationTimestamp="2025-10-08 15:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:42.429985486 +0000 UTC m=+1471.783900407" watchObservedRunningTime="2025-10-08 15:37:42.441760733 +0000 UTC m=+1471.795675654" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.479597 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.479572067 podStartE2EDuration="2.479572067s" podCreationTimestamp="2025-10-08 15:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:42.475719884 +0000 UTC m=+1471.829634795" watchObservedRunningTime="2025-10-08 15:37:42.479572067 +0000 UTC m=+1471.833487008" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.629082 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-766ff48cb5-86ctl"] Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.631986 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.653202 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766ff48cb5-86ctl"] Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.796220 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7ctk\" (UniqueName: \"kubernetes.io/projected/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-kube-api-access-p7ctk\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.796715 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-svc\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.797023 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-sb\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.797338 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-swift-storage-0\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.797674 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-nb\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.798045 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-config\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.899448 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-config\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.900170 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7ctk\" (UniqueName: \"kubernetes.io/projected/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-kube-api-access-p7ctk\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.900279 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-svc\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.900529 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-config\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.900970 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-svc\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.901057 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-sb\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.901255 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-swift-storage-0\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.901348 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-nb\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.901990 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-nb\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.902547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-sb\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.903776 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-swift-storage-0\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.923718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7ctk\" (UniqueName: \"kubernetes.io/projected/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-kube-api-access-p7ctk\") pod \"dnsmasq-dns-766ff48cb5-86ctl\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:42 crc kubenswrapper[4945]: I1008 15:37:42.958640 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:43 crc kubenswrapper[4945]: I1008 15:37:43.295442 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766ff48cb5-86ctl"] Oct 08 15:37:43 crc kubenswrapper[4945]: I1008 15:37:43.417877 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" event={"ID":"6811fe37-bd5a-4a52-91a6-91be2ad24e4c","Type":"ContainerStarted","Data":"2599d2af091cfdc1f035a278b3700769e94adda07526463d4073da6c96419fc8"} Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.429082 4945 generic.go:334] "Generic (PLEG): container finished" podID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerID="30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c" exitCode=0 Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.429193 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" event={"ID":"6811fe37-bd5a-4a52-91a6-91be2ad24e4c","Type":"ContainerDied","Data":"30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c"} Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.939574 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.940354 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-central-agent" containerID="cri-o://c9a8c80ddd1aa76aa4e1b46b53c12afafc8bfce467023778ceb205543f315391" gracePeriod=30 Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.940394 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="sg-core" containerID="cri-o://a7111a26e0e8b428da8f322c818a89bb83840cdbf61aeb83d9a681262b4671bf" gracePeriod=30 Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.940436 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="proxy-httpd" containerID="cri-o://6852350f982566884c1e617f95195f371ca613a68c032dbbb963da9aae8b0dfb" gracePeriod=30 Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.940479 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-notification-agent" containerID="cri-o://00a9b58478c062a2190516333e7ba8ce54b4580668f5d309fa459da9eb5f6743" gracePeriod=30 Oct 08 15:37:44 crc kubenswrapper[4945]: I1008 15:37:44.951349 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.212:3000/\": EOF" Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.074630 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.441881 4945 generic.go:334] "Generic (PLEG): container finished" podID="0908392e-ee1c-4220-a893-0da9b12f813b" containerID="6852350f982566884c1e617f95195f371ca613a68c032dbbb963da9aae8b0dfb" exitCode=0 Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.441917 4945 generic.go:334] "Generic (PLEG): container finished" podID="0908392e-ee1c-4220-a893-0da9b12f813b" containerID="a7111a26e0e8b428da8f322c818a89bb83840cdbf61aeb83d9a681262b4671bf" exitCode=2 Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.441927 4945 generic.go:334] "Generic (PLEG): container finished" podID="0908392e-ee1c-4220-a893-0da9b12f813b" containerID="c9a8c80ddd1aa76aa4e1b46b53c12afafc8bfce467023778ceb205543f315391" exitCode=0 Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.441974 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerDied","Data":"6852350f982566884c1e617f95195f371ca613a68c032dbbb963da9aae8b0dfb"} Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.442006 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerDied","Data":"a7111a26e0e8b428da8f322c818a89bb83840cdbf61aeb83d9a681262b4671bf"} Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.442017 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerDied","Data":"c9a8c80ddd1aa76aa4e1b46b53c12afafc8bfce467023778ceb205543f315391"} Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.444549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" event={"ID":"6811fe37-bd5a-4a52-91a6-91be2ad24e4c","Type":"ContainerStarted","Data":"a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d"} Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.444698 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-log" containerID="cri-o://be33f8dcd7a09c8c2fbf288b6b0a761ef7ac831c2f2ec6007c8d23a136d9b6cd" gracePeriod=30 Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.444779 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-api" containerID="cri-o://2cf7a2a1b470261f3b61e2c6ea3e6c1fedca423033cd911e3b3b8bb6e39eeddc" gracePeriod=30 Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.479501 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" podStartSLOduration=3.479478596 podStartE2EDuration="3.479478596s" podCreationTimestamp="2025-10-08 15:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:45.471341347 +0000 UTC m=+1474.825256248" watchObservedRunningTime="2025-10-08 15:37:45.479478596 +0000 UTC m=+1474.833393497" Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.809202 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 15:37:45 crc kubenswrapper[4945]: I1008 15:37:45.809681 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.087283 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.458488 4945 generic.go:334] "Generic (PLEG): container finished" podID="26f7b93f-2750-47ef-a855-c3479a38759a" containerID="2cf7a2a1b470261f3b61e2c6ea3e6c1fedca423033cd911e3b3b8bb6e39eeddc" exitCode=0 Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.458520 4945 generic.go:334] "Generic (PLEG): container finished" podID="26f7b93f-2750-47ef-a855-c3479a38759a" containerID="be33f8dcd7a09c8c2fbf288b6b0a761ef7ac831c2f2ec6007c8d23a136d9b6cd" exitCode=143 Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.458593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f7b93f-2750-47ef-a855-c3479a38759a","Type":"ContainerDied","Data":"2cf7a2a1b470261f3b61e2c6ea3e6c1fedca423033cd911e3b3b8bb6e39eeddc"} Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.458653 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f7b93f-2750-47ef-a855-c3479a38759a","Type":"ContainerDied","Data":"be33f8dcd7a09c8c2fbf288b6b0a761ef7ac831c2f2ec6007c8d23a136d9b6cd"} Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.458753 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.695442 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.212:3000/\": dial tcp 10.217.0.212:3000: connect: connection refused" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.744769 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.891304 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-combined-ca-bundle\") pod \"26f7b93f-2750-47ef-a855-c3479a38759a\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.891419 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-config-data\") pod \"26f7b93f-2750-47ef-a855-c3479a38759a\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.891656 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l68tc\" (UniqueName: \"kubernetes.io/projected/26f7b93f-2750-47ef-a855-c3479a38759a-kube-api-access-l68tc\") pod \"26f7b93f-2750-47ef-a855-c3479a38759a\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.891691 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f7b93f-2750-47ef-a855-c3479a38759a-logs\") pod \"26f7b93f-2750-47ef-a855-c3479a38759a\" (UID: \"26f7b93f-2750-47ef-a855-c3479a38759a\") " Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.892699 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f7b93f-2750-47ef-a855-c3479a38759a-logs" (OuterVolumeSpecName: "logs") pod "26f7b93f-2750-47ef-a855-c3479a38759a" (UID: "26f7b93f-2750-47ef-a855-c3479a38759a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.898908 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f7b93f-2750-47ef-a855-c3479a38759a-kube-api-access-l68tc" (OuterVolumeSpecName: "kube-api-access-l68tc") pod "26f7b93f-2750-47ef-a855-c3479a38759a" (UID: "26f7b93f-2750-47ef-a855-c3479a38759a"). InnerVolumeSpecName "kube-api-access-l68tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.934980 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26f7b93f-2750-47ef-a855-c3479a38759a" (UID: "26f7b93f-2750-47ef-a855-c3479a38759a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.951662 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-config-data" (OuterVolumeSpecName: "config-data") pod "26f7b93f-2750-47ef-a855-c3479a38759a" (UID: "26f7b93f-2750-47ef-a855-c3479a38759a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.993720 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l68tc\" (UniqueName: \"kubernetes.io/projected/26f7b93f-2750-47ef-a855-c3479a38759a-kube-api-access-l68tc\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.993765 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26f7b93f-2750-47ef-a855-c3479a38759a-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.993779 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:46 crc kubenswrapper[4945]: I1008 15:37:46.993790 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26f7b93f-2750-47ef-a855-c3479a38759a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.478541 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.478539 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"26f7b93f-2750-47ef-a855-c3479a38759a","Type":"ContainerDied","Data":"8e93f66f98f9155e3b2d7fb144bfc9de0ca354da679fa26df349db5324a63501"} Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.478987 4945 scope.go:117] "RemoveContainer" containerID="2cf7a2a1b470261f3b61e2c6ea3e6c1fedca423033cd911e3b3b8bb6e39eeddc" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.511892 4945 scope.go:117] "RemoveContainer" containerID="be33f8dcd7a09c8c2fbf288b6b0a761ef7ac831c2f2ec6007c8d23a136d9b6cd" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.516645 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.534129 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.548253 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:47 crc kubenswrapper[4945]: E1008 15:37:47.548640 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-api" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.548655 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-api" Oct 08 15:37:47 crc kubenswrapper[4945]: E1008 15:37:47.548686 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-log" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.548692 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-log" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.548909 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-log" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.548925 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" containerName="nova-api-api" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.549941 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.552524 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.566478 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.566690 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.584506 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.710705 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l9zj\" (UniqueName: \"kubernetes.io/projected/6fed2c6e-b5c7-4128-9aef-a310632ca057-kube-api-access-6l9zj\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.710792 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-config-data\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.711055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-public-tls-certs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.711174 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.711282 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fed2c6e-b5c7-4128-9aef-a310632ca057-logs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.711424 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.813351 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l9zj\" (UniqueName: \"kubernetes.io/projected/6fed2c6e-b5c7-4128-9aef-a310632ca057-kube-api-access-6l9zj\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.813409 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-config-data\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.813492 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-public-tls-certs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.813516 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.813554 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fed2c6e-b5c7-4128-9aef-a310632ca057-logs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.813606 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.817905 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fed2c6e-b5c7-4128-9aef-a310632ca057-logs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.818993 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-public-tls-certs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.821345 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-config-data\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.821487 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.834999 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.836930 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l9zj\" (UniqueName: \"kubernetes.io/projected/6fed2c6e-b5c7-4128-9aef-a310632ca057-kube-api-access-6l9zj\") pod \"nova-api-0\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " pod="openstack/nova-api-0" Oct 08 15:37:47 crc kubenswrapper[4945]: I1008 15:37:47.881368 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.041256 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f7b93f-2750-47ef-a855-c3479a38759a" path="/var/lib/kubelet/pods/26f7b93f-2750-47ef-a855-c3479a38759a/volumes" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.345507 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.491748 4945 generic.go:334] "Generic (PLEG): container finished" podID="0908392e-ee1c-4220-a893-0da9b12f813b" containerID="00a9b58478c062a2190516333e7ba8ce54b4580668f5d309fa459da9eb5f6743" exitCode=0 Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.491825 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerDied","Data":"00a9b58478c062a2190516333e7ba8ce54b4580668f5d309fa459da9eb5f6743"} Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.493522 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fed2c6e-b5c7-4128-9aef-a310632ca057","Type":"ContainerStarted","Data":"8caa83dfcfd036670b1df0846560a6df7cf4813280ca38934a2f1a26ef73fb6c"} Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.539744 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636207 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-combined-ca-bundle\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636268 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-scripts\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636383 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-run-httpd\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636426 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-sg-core-conf-yaml\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636479 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-config-data\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-log-httpd\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636674 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx476\" (UniqueName: \"kubernetes.io/projected/0908392e-ee1c-4220-a893-0da9b12f813b-kube-api-access-tx476\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.636739 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-ceilometer-tls-certs\") pod \"0908392e-ee1c-4220-a893-0da9b12f813b\" (UID: \"0908392e-ee1c-4220-a893-0da9b12f813b\") " Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.637154 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.637365 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.637533 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.642028 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-scripts" (OuterVolumeSpecName: "scripts") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.642188 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0908392e-ee1c-4220-a893-0da9b12f813b-kube-api-access-tx476" (OuterVolumeSpecName: "kube-api-access-tx476") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "kube-api-access-tx476". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.673332 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.700122 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.736905 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.739373 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx476\" (UniqueName: \"kubernetes.io/projected/0908392e-ee1c-4220-a893-0da9b12f813b-kube-api-access-tx476\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.739440 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.739454 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.739466 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.739480 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0908392e-ee1c-4220-a893-0da9b12f813b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.739492 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.760749 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-config-data" (OuterVolumeSpecName: "config-data") pod "0908392e-ee1c-4220-a893-0da9b12f813b" (UID: "0908392e-ee1c-4220-a893-0da9b12f813b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:48 crc kubenswrapper[4945]: I1008 15:37:48.841666 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0908392e-ee1c-4220-a893-0da9b12f813b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.505751 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fed2c6e-b5c7-4128-9aef-a310632ca057","Type":"ContainerStarted","Data":"e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95"} Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.506328 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fed2c6e-b5c7-4128-9aef-a310632ca057","Type":"ContainerStarted","Data":"e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea"} Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.508635 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0908392e-ee1c-4220-a893-0da9b12f813b","Type":"ContainerDied","Data":"efa1d3aa7a6a3a224760ed859e2e059a032cf219e36ccabb435166097da04fdc"} Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.508668 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.508688 4945 scope.go:117] "RemoveContainer" containerID="6852350f982566884c1e617f95195f371ca613a68c032dbbb963da9aae8b0dfb" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.540364 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5401509300000003 podStartE2EDuration="2.54015093s" podCreationTimestamp="2025-10-08 15:37:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:49.527196323 +0000 UTC m=+1478.881111244" watchObservedRunningTime="2025-10-08 15:37:49.54015093 +0000 UTC m=+1478.894065831" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.550709 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.552410 4945 scope.go:117] "RemoveContainer" containerID="a7111a26e0e8b428da8f322c818a89bb83840cdbf61aeb83d9a681262b4671bf" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.558753 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.574672 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:49 crc kubenswrapper[4945]: E1008 15:37:49.575316 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="proxy-httpd" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.575388 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="proxy-httpd" Oct 08 15:37:49 crc kubenswrapper[4945]: E1008 15:37:49.575476 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-notification-agent" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.575524 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-notification-agent" Oct 08 15:37:49 crc kubenswrapper[4945]: E1008 15:37:49.575576 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-central-agent" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.575651 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-central-agent" Oct 08 15:37:49 crc kubenswrapper[4945]: E1008 15:37:49.575710 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="sg-core" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.575770 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="sg-core" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.578487 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="proxy-httpd" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.578658 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="sg-core" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.578738 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-central-agent" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.578790 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" containerName="ceilometer-notification-agent" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.581136 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.583994 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.584037 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.584241 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.592962 4945 scope.go:117] "RemoveContainer" containerID="00a9b58478c062a2190516333e7ba8ce54b4580668f5d309fa459da9eb5f6743" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.599544 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.647090 4945 scope.go:117] "RemoveContainer" containerID="c9a8c80ddd1aa76aa4e1b46b53c12afafc8bfce467023778ceb205543f315391" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.758937 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-scripts\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.759075 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.759133 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.759264 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m5dx\" (UniqueName: \"kubernetes.io/projected/950bbdee-1a42-4890-b678-abd7e6a65402-kube-api-access-6m5dx\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.759372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/950bbdee-1a42-4890-b678-abd7e6a65402-log-httpd\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.759438 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-config-data\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.759483 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/950bbdee-1a42-4890-b678-abd7e6a65402-run-httpd\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.759579 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:49 crc kubenswrapper[4945]: I1008 15:37:49.860947 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m5dx\" (UniqueName: \"kubernetes.io/projected/950bbdee-1a42-4890-b678-abd7e6a65402-kube-api-access-6m5dx\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861060 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/950bbdee-1a42-4890-b678-abd7e6a65402-log-httpd\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861087 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-config-data\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861130 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/950bbdee-1a42-4890-b678-abd7e6a65402-run-httpd\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861214 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-scripts\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861280 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.861598 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/950bbdee-1a42-4890-b678-abd7e6a65402-log-httpd\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.862155 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/950bbdee-1a42-4890-b678-abd7e6a65402-run-httpd\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.866629 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.871958 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.872319 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.876033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-config-data\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.878451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/950bbdee-1a42-4890-b678-abd7e6a65402-scripts\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.879266 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m5dx\" (UniqueName: \"kubernetes.io/projected/950bbdee-1a42-4890-b678-abd7e6a65402-kube-api-access-6m5dx\") pod \"ceilometer-0\" (UID: \"950bbdee-1a42-4890-b678-abd7e6a65402\") " pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:49.917697 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:50.037894 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0908392e-ee1c-4220-a893-0da9b12f813b" path="/var/lib/kubelet/pods/0908392e-ee1c-4220-a893-0da9b12f813b/volumes" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:50.796626 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:50.809466 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 15:37:50 crc kubenswrapper[4945]: I1008 15:37:50.810763 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.087207 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.104355 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.533307 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"950bbdee-1a42-4890-b678-abd7e6a65402","Type":"ContainerStarted","Data":"4f833128903af06e033f75fed020439681a5918fb1d9c5e187091bca090d03f2"} Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.540627 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"950bbdee-1a42-4890-b678-abd7e6a65402","Type":"ContainerStarted","Data":"487662b779e75c9c4412ba4e16ee0a617ace43aecdd08b340e8cbcf8ba312b54"} Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.540690 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"950bbdee-1a42-4890-b678-abd7e6a65402","Type":"ContainerStarted","Data":"d832de09a4165b5fc09dd69c96c00084290f35f17ffbe079a65084b411fe4173"} Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.554406 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.725400 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-xk6lb"] Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.726752 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.740572 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.740766 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.742775 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xk6lb"] Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.800184 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-scripts\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.800258 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.800287 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lqw2\" (UniqueName: \"kubernetes.io/projected/24aa0642-d709-4118-ad55-fd216646ecc2-kube-api-access-2lqw2\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.800357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-config-data\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.815358 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.217:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.827232 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.217:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.902320 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-config-data\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.902497 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-scripts\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.902652 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.902754 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lqw2\" (UniqueName: \"kubernetes.io/projected/24aa0642-d709-4118-ad55-fd216646ecc2-kube-api-access-2lqw2\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.910449 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-scripts\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.911356 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.916657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-config-data\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:51 crc kubenswrapper[4945]: I1008 15:37:51.928821 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lqw2\" (UniqueName: \"kubernetes.io/projected/24aa0642-d709-4118-ad55-fd216646ecc2-kube-api-access-2lqw2\") pod \"nova-cell1-cell-mapping-xk6lb\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:52 crc kubenswrapper[4945]: I1008 15:37:52.060773 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:37:52 crc kubenswrapper[4945]: I1008 15:37:52.569986 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"950bbdee-1a42-4890-b678-abd7e6a65402","Type":"ContainerStarted","Data":"97247bc0768d9a62a244f1d83ec9c88ea478e196920dfb803350cbacdc15076b"} Oct 08 15:37:52 crc kubenswrapper[4945]: I1008 15:37:52.578318 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xk6lb"] Oct 08 15:37:52 crc kubenswrapper[4945]: W1008 15:37:52.589254 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24aa0642_d709_4118_ad55_fd216646ecc2.slice/crio-c6fb82c46ba637cc73cb893f18089018fd9983575a0700ee08f68c866ef01326 WatchSource:0}: Error finding container c6fb82c46ba637cc73cb893f18089018fd9983575a0700ee08f68c866ef01326: Status 404 returned error can't find the container with id c6fb82c46ba637cc73cb893f18089018fd9983575a0700ee08f68c866ef01326 Oct 08 15:37:52 crc kubenswrapper[4945]: I1008 15:37:52.960377 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.034263 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84d796b5f9-vqkd7"] Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.034542 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" podUID="b402b060-866f-40db-a058-74a48be28f64" containerName="dnsmasq-dns" containerID="cri-o://ee3647dd523a3881a0a0add10110fa5a1b57bca8f3ea7e735eff6f6a98d93b03" gracePeriod=10 Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.581420 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xk6lb" event={"ID":"24aa0642-d709-4118-ad55-fd216646ecc2","Type":"ContainerStarted","Data":"f0eacbb9d4443ec00ff22d374cb725dbcce021dfc4d84790858a190abf0810d0"} Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.581704 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xk6lb" event={"ID":"24aa0642-d709-4118-ad55-fd216646ecc2","Type":"ContainerStarted","Data":"c6fb82c46ba637cc73cb893f18089018fd9983575a0700ee08f68c866ef01326"} Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.584830 4945 generic.go:334] "Generic (PLEG): container finished" podID="b402b060-866f-40db-a058-74a48be28f64" containerID="ee3647dd523a3881a0a0add10110fa5a1b57bca8f3ea7e735eff6f6a98d93b03" exitCode=0 Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.584884 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" event={"ID":"b402b060-866f-40db-a058-74a48be28f64","Type":"ContainerDied","Data":"ee3647dd523a3881a0a0add10110fa5a1b57bca8f3ea7e735eff6f6a98d93b03"} Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.584916 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" event={"ID":"b402b060-866f-40db-a058-74a48be28f64","Type":"ContainerDied","Data":"440a722f4e385942156a2ec19a7b0e1a1e47c4d473edbe1c9d33f2f3f7c97bc3"} Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.584929 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="440a722f4e385942156a2ec19a7b0e1a1e47c4d473edbe1c9d33f2f3f7c97bc3" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.605244 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-xk6lb" podStartSLOduration=2.60522213 podStartE2EDuration="2.60522213s" podCreationTimestamp="2025-10-08 15:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:37:53.59950241 +0000 UTC m=+1482.953417331" watchObservedRunningTime="2025-10-08 15:37:53.60522213 +0000 UTC m=+1482.959137031" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.658819 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.756076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-swift-storage-0\") pod \"b402b060-866f-40db-a058-74a48be28f64\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.757007 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-svc\") pod \"b402b060-866f-40db-a058-74a48be28f64\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.757063 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb\") pod \"b402b060-866f-40db-a058-74a48be28f64\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.757159 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-sb\") pod \"b402b060-866f-40db-a058-74a48be28f64\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.757211 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh6kz\" (UniqueName: \"kubernetes.io/projected/b402b060-866f-40db-a058-74a48be28f64-kube-api-access-qh6kz\") pod \"b402b060-866f-40db-a058-74a48be28f64\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.757277 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-config\") pod \"b402b060-866f-40db-a058-74a48be28f64\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.792911 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b402b060-866f-40db-a058-74a48be28f64-kube-api-access-qh6kz" (OuterVolumeSpecName: "kube-api-access-qh6kz") pod "b402b060-866f-40db-a058-74a48be28f64" (UID: "b402b060-866f-40db-a058-74a48be28f64"). InnerVolumeSpecName "kube-api-access-qh6kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.834940 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-config" (OuterVolumeSpecName: "config") pod "b402b060-866f-40db-a058-74a48be28f64" (UID: "b402b060-866f-40db-a058-74a48be28f64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.850314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b402b060-866f-40db-a058-74a48be28f64" (UID: "b402b060-866f-40db-a058-74a48be28f64"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.856819 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b402b060-866f-40db-a058-74a48be28f64" (UID: "b402b060-866f-40db-a058-74a48be28f64"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.861679 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b402b060-866f-40db-a058-74a48be28f64" (UID: "b402b060-866f-40db-a058-74a48be28f64"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.861832 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb\") pod \"b402b060-866f-40db-a058-74a48be28f64\" (UID: \"b402b060-866f-40db-a058-74a48be28f64\") " Oct 08 15:37:53 crc kubenswrapper[4945]: W1008 15:37:53.862131 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b402b060-866f-40db-a058-74a48be28f64/volumes/kubernetes.io~configmap/ovsdbserver-nb Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.862157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b402b060-866f-40db-a058-74a48be28f64" (UID: "b402b060-866f-40db-a058-74a48be28f64"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.862479 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.862494 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.862504 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.862512 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh6kz\" (UniqueName: \"kubernetes.io/projected/b402b060-866f-40db-a058-74a48be28f64-kube-api-access-qh6kz\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.862522 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.870485 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b402b060-866f-40db-a058-74a48be28f64" (UID: "b402b060-866f-40db-a058-74a48be28f64"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:53 crc kubenswrapper[4945]: I1008 15:37:53.964714 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b402b060-866f-40db-a058-74a48be28f64-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:54 crc kubenswrapper[4945]: I1008 15:37:54.602359 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"950bbdee-1a42-4890-b678-abd7e6a65402","Type":"ContainerStarted","Data":"dbbbb48643f4b223ae4eb853d6117119de73c5addc14b2844bf8e1806a52ad06"} Oct 08 15:37:54 crc kubenswrapper[4945]: I1008 15:37:54.602430 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d796b5f9-vqkd7" Oct 08 15:37:54 crc kubenswrapper[4945]: I1008 15:37:54.632209 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.355724955 podStartE2EDuration="5.63219075s" podCreationTimestamp="2025-10-08 15:37:49 +0000 UTC" firstStartedPulling="2025-10-08 15:37:50.813680783 +0000 UTC m=+1480.167595704" lastFinishedPulling="2025-10-08 15:37:54.090146598 +0000 UTC m=+1483.444061499" observedRunningTime="2025-10-08 15:37:54.623221451 +0000 UTC m=+1483.977136352" watchObservedRunningTime="2025-10-08 15:37:54.63219075 +0000 UTC m=+1483.986105651" Oct 08 15:37:54 crc kubenswrapper[4945]: I1008 15:37:54.656210 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84d796b5f9-vqkd7"] Oct 08 15:37:54 crc kubenswrapper[4945]: I1008 15:37:54.666405 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84d796b5f9-vqkd7"] Oct 08 15:37:55 crc kubenswrapper[4945]: I1008 15:37:55.611637 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 15:37:56 crc kubenswrapper[4945]: I1008 15:37:56.042803 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b402b060-866f-40db-a058-74a48be28f64" path="/var/lib/kubelet/pods/b402b060-866f-40db-a058-74a48be28f64/volumes" Oct 08 15:37:57 crc kubenswrapper[4945]: I1008 15:37:57.881686 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:37:57 crc kubenswrapper[4945]: I1008 15:37:57.881974 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:37:58 crc kubenswrapper[4945]: I1008 15:37:58.643916 4945 generic.go:334] "Generic (PLEG): container finished" podID="24aa0642-d709-4118-ad55-fd216646ecc2" containerID="f0eacbb9d4443ec00ff22d374cb725dbcce021dfc4d84790858a190abf0810d0" exitCode=0 Oct 08 15:37:58 crc kubenswrapper[4945]: I1008 15:37:58.644302 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xk6lb" event={"ID":"24aa0642-d709-4118-ad55-fd216646ecc2","Type":"ContainerDied","Data":"f0eacbb9d4443ec00ff22d374cb725dbcce021dfc4d84790858a190abf0810d0"} Oct 08 15:37:58 crc kubenswrapper[4945]: I1008 15:37:58.895460 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.219:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:37:58 crc kubenswrapper[4945]: I1008 15:37:58.895460 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.219:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.067918 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.208598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-combined-ca-bundle\") pod \"24aa0642-d709-4118-ad55-fd216646ecc2\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.208938 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lqw2\" (UniqueName: \"kubernetes.io/projected/24aa0642-d709-4118-ad55-fd216646ecc2-kube-api-access-2lqw2\") pod \"24aa0642-d709-4118-ad55-fd216646ecc2\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.209071 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-scripts\") pod \"24aa0642-d709-4118-ad55-fd216646ecc2\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.209197 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-config-data\") pod \"24aa0642-d709-4118-ad55-fd216646ecc2\" (UID: \"24aa0642-d709-4118-ad55-fd216646ecc2\") " Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.213728 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24aa0642-d709-4118-ad55-fd216646ecc2-kube-api-access-2lqw2" (OuterVolumeSpecName: "kube-api-access-2lqw2") pod "24aa0642-d709-4118-ad55-fd216646ecc2" (UID: "24aa0642-d709-4118-ad55-fd216646ecc2"). InnerVolumeSpecName "kube-api-access-2lqw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.214211 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-scripts" (OuterVolumeSpecName: "scripts") pod "24aa0642-d709-4118-ad55-fd216646ecc2" (UID: "24aa0642-d709-4118-ad55-fd216646ecc2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.239403 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24aa0642-d709-4118-ad55-fd216646ecc2" (UID: "24aa0642-d709-4118-ad55-fd216646ecc2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.256968 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-config-data" (OuterVolumeSpecName: "config-data") pod "24aa0642-d709-4118-ad55-fd216646ecc2" (UID: "24aa0642-d709-4118-ad55-fd216646ecc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.314466 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.314511 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.314529 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lqw2\" (UniqueName: \"kubernetes.io/projected/24aa0642-d709-4118-ad55-fd216646ecc2-kube-api-access-2lqw2\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.314540 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24aa0642-d709-4118-ad55-fd216646ecc2-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.667208 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xk6lb" event={"ID":"24aa0642-d709-4118-ad55-fd216646ecc2","Type":"ContainerDied","Data":"c6fb82c46ba637cc73cb893f18089018fd9983575a0700ee08f68c866ef01326"} Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.667244 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6fb82c46ba637cc73cb893f18089018fd9983575a0700ee08f68c866ef01326" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.667299 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xk6lb" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.813754 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.815159 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.837883 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.885098 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.885367 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-log" containerID="cri-o://e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea" gracePeriod=30 Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.885508 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-api" containerID="cri-o://e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95" gracePeriod=30 Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.925651 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.925920 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b3a9136d-3567-4523-97bb-56936b7f6ae4" containerName="nova-scheduler-scheduler" containerID="cri-o://b9446e663432cd61b397f3728603ba4c2c67baddfdc63924a70e6024bf3f3e21" gracePeriod=30 Oct 08 15:38:00 crc kubenswrapper[4945]: I1008 15:38:00.964096 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:38:01 crc kubenswrapper[4945]: I1008 15:38:01.681491 4945 generic.go:334] "Generic (PLEG): container finished" podID="b3a9136d-3567-4523-97bb-56936b7f6ae4" containerID="b9446e663432cd61b397f3728603ba4c2c67baddfdc63924a70e6024bf3f3e21" exitCode=0 Oct 08 15:38:01 crc kubenswrapper[4945]: I1008 15:38:01.681567 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3a9136d-3567-4523-97bb-56936b7f6ae4","Type":"ContainerDied","Data":"b9446e663432cd61b397f3728603ba4c2c67baddfdc63924a70e6024bf3f3e21"} Oct 08 15:38:01 crc kubenswrapper[4945]: I1008 15:38:01.684842 4945 generic.go:334] "Generic (PLEG): container finished" podID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerID="e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea" exitCode=143 Oct 08 15:38:01 crc kubenswrapper[4945]: I1008 15:38:01.684910 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fed2c6e-b5c7-4128-9aef-a310632ca057","Type":"ContainerDied","Data":"e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea"} Oct 08 15:38:01 crc kubenswrapper[4945]: I1008 15:38:01.690399 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 15:38:01 crc kubenswrapper[4945]: I1008 15:38:01.941804 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.053325 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-combined-ca-bundle\") pod \"b3a9136d-3567-4523-97bb-56936b7f6ae4\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.053886 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8pkp\" (UniqueName: \"kubernetes.io/projected/b3a9136d-3567-4523-97bb-56936b7f6ae4-kube-api-access-g8pkp\") pod \"b3a9136d-3567-4523-97bb-56936b7f6ae4\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.054018 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-config-data\") pod \"b3a9136d-3567-4523-97bb-56936b7f6ae4\" (UID: \"b3a9136d-3567-4523-97bb-56936b7f6ae4\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.061370 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3a9136d-3567-4523-97bb-56936b7f6ae4-kube-api-access-g8pkp" (OuterVolumeSpecName: "kube-api-access-g8pkp") pod "b3a9136d-3567-4523-97bb-56936b7f6ae4" (UID: "b3a9136d-3567-4523-97bb-56936b7f6ae4"). InnerVolumeSpecName "kube-api-access-g8pkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.081923 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3a9136d-3567-4523-97bb-56936b7f6ae4" (UID: "b3a9136d-3567-4523-97bb-56936b7f6ae4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.100370 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-config-data" (OuterVolumeSpecName: "config-data") pod "b3a9136d-3567-4523-97bb-56936b7f6ae4" (UID: "b3a9136d-3567-4523-97bb-56936b7f6ae4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.156540 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8pkp\" (UniqueName: \"kubernetes.io/projected/b3a9136d-3567-4523-97bb-56936b7f6ae4-kube-api-access-g8pkp\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.156568 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.156580 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a9136d-3567-4523-97bb-56936b7f6ae4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.208778 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.359225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-internal-tls-certs\") pod \"6fed2c6e-b5c7-4128-9aef-a310632ca057\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.359292 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-combined-ca-bundle\") pod \"6fed2c6e-b5c7-4128-9aef-a310632ca057\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.359449 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-public-tls-certs\") pod \"6fed2c6e-b5c7-4128-9aef-a310632ca057\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.359495 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-config-data\") pod \"6fed2c6e-b5c7-4128-9aef-a310632ca057\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.359534 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fed2c6e-b5c7-4128-9aef-a310632ca057-logs\") pod \"6fed2c6e-b5c7-4128-9aef-a310632ca057\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.359665 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l9zj\" (UniqueName: \"kubernetes.io/projected/6fed2c6e-b5c7-4128-9aef-a310632ca057-kube-api-access-6l9zj\") pod \"6fed2c6e-b5c7-4128-9aef-a310632ca057\" (UID: \"6fed2c6e-b5c7-4128-9aef-a310632ca057\") " Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.360640 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fed2c6e-b5c7-4128-9aef-a310632ca057-logs" (OuterVolumeSpecName: "logs") pod "6fed2c6e-b5c7-4128-9aef-a310632ca057" (UID: "6fed2c6e-b5c7-4128-9aef-a310632ca057"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.366582 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fed2c6e-b5c7-4128-9aef-a310632ca057-kube-api-access-6l9zj" (OuterVolumeSpecName: "kube-api-access-6l9zj") pod "6fed2c6e-b5c7-4128-9aef-a310632ca057" (UID: "6fed2c6e-b5c7-4128-9aef-a310632ca057"). InnerVolumeSpecName "kube-api-access-6l9zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.396926 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-config-data" (OuterVolumeSpecName: "config-data") pod "6fed2c6e-b5c7-4128-9aef-a310632ca057" (UID: "6fed2c6e-b5c7-4128-9aef-a310632ca057"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.421649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fed2c6e-b5c7-4128-9aef-a310632ca057" (UID: "6fed2c6e-b5c7-4128-9aef-a310632ca057"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.423792 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6fed2c6e-b5c7-4128-9aef-a310632ca057" (UID: "6fed2c6e-b5c7-4128-9aef-a310632ca057"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.427357 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6fed2c6e-b5c7-4128-9aef-a310632ca057" (UID: "6fed2c6e-b5c7-4128-9aef-a310632ca057"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.461527 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.461560 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.461583 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.461597 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fed2c6e-b5c7-4128-9aef-a310632ca057-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.461607 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l9zj\" (UniqueName: \"kubernetes.io/projected/6fed2c6e-b5c7-4128-9aef-a310632ca057-kube-api-access-6l9zj\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.461621 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fed2c6e-b5c7-4128-9aef-a310632ca057-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.695596 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3a9136d-3567-4523-97bb-56936b7f6ae4","Type":"ContainerDied","Data":"4cd71deaaa5b3d2bae801c7a8d9a17aa1ed7383230779fb26ff09cdd40ee9799"} Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.695604 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.695651 4945 scope.go:117] "RemoveContainer" containerID="b9446e663432cd61b397f3728603ba4c2c67baddfdc63924a70e6024bf3f3e21" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.699229 4945 generic.go:334] "Generic (PLEG): container finished" podID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerID="e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95" exitCode=0 Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.699509 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-log" containerID="cri-o://4711b6a17f7f102d9385a94f060dcdb2664bb77fd86d1db33c0cec5823c0ee92" gracePeriod=30 Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.699808 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.700203 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fed2c6e-b5c7-4128-9aef-a310632ca057","Type":"ContainerDied","Data":"e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95"} Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.700245 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fed2c6e-b5c7-4128-9aef-a310632ca057","Type":"ContainerDied","Data":"8caa83dfcfd036670b1df0846560a6df7cf4813280ca38934a2f1a26ef73fb6c"} Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.700276 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-metadata" containerID="cri-o://1fbb94a8d8f61a0b6cef4a5fdb687b9157cf516718378738b1a4703dec50900b" gracePeriod=30 Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.732876 4945 scope.go:117] "RemoveContainer" containerID="e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.737991 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.756308 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.767699 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.779310 4945 scope.go:117] "RemoveContainer" containerID="e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.784883 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.798816 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.799761 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24aa0642-d709-4118-ad55-fd216646ecc2" containerName="nova-manage" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.800041 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="24aa0642-d709-4118-ad55-fd216646ecc2" containerName="nova-manage" Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.800189 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-log" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.800276 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-log" Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.800339 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b402b060-866f-40db-a058-74a48be28f64" containerName="init" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.800387 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b402b060-866f-40db-a058-74a48be28f64" containerName="init" Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.800436 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b402b060-866f-40db-a058-74a48be28f64" containerName="dnsmasq-dns" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.800489 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b402b060-866f-40db-a058-74a48be28f64" containerName="dnsmasq-dns" Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.800560 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-api" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.800607 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-api" Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.800678 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3a9136d-3567-4523-97bb-56936b7f6ae4" containerName="nova-scheduler-scheduler" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.800741 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3a9136d-3567-4523-97bb-56936b7f6ae4" containerName="nova-scheduler-scheduler" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.801072 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-log" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.801176 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="24aa0642-d709-4118-ad55-fd216646ecc2" containerName="nova-manage" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.801239 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3a9136d-3567-4523-97bb-56936b7f6ae4" containerName="nova-scheduler-scheduler" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.801356 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" containerName="nova-api-api" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.801434 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b402b060-866f-40db-a058-74a48be28f64" containerName="dnsmasq-dns" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.802231 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.804982 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.808179 4945 scope.go:117] "RemoveContainer" containerID="e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.811090 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.820717 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.820926 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95\": container with ID starting with e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95 not found: ID does not exist" containerID="e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.820981 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95"} err="failed to get container status \"e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95\": rpc error: code = NotFound desc = could not find container \"e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95\": container with ID starting with e8c469e17ff244567cdeef48484467a015748324134e1bd973c9ac3f961dee95 not found: ID does not exist" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.821011 4945 scope.go:117] "RemoveContainer" containerID="e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea" Oct 08 15:38:02 crc kubenswrapper[4945]: E1008 15:38:02.821534 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea\": container with ID starting with e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea not found: ID does not exist" containerID="e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.822493 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea"} err="failed to get container status \"e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea\": rpc error: code = NotFound desc = could not find container \"e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea\": container with ID starting with e547ee89faa2965eb1aa218d0a5c620766185e2bc9b1fff9017df1423f8910ea not found: ID does not exist" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.826472 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.828948 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.830519 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.830621 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.832081 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.972960 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17f16f2-5d8f-4178-a6b5-879444c7b262-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973006 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8e508c6-7afc-4ae0-84bb-48dece116b1d-logs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973111 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mldtj\" (UniqueName: \"kubernetes.io/projected/c17f16f2-5d8f-4178-a6b5-879444c7b262-kube-api-access-mldtj\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973285 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw865\" (UniqueName: \"kubernetes.io/projected/e8e508c6-7afc-4ae0-84bb-48dece116b1d-kube-api-access-zw865\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973320 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973339 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17f16f2-5d8f-4178-a6b5-879444c7b262-config-data\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973374 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-public-tls-certs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973393 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:02 crc kubenswrapper[4945]: I1008 15:38:02.973486 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-config-data\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076003 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076077 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-config-data\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076146 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17f16f2-5d8f-4178-a6b5-879444c7b262-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076172 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8e508c6-7afc-4ae0-84bb-48dece116b1d-logs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076375 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mldtj\" (UniqueName: \"kubernetes.io/projected/c17f16f2-5d8f-4178-a6b5-879444c7b262-kube-api-access-mldtj\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076477 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw865\" (UniqueName: \"kubernetes.io/projected/e8e508c6-7afc-4ae0-84bb-48dece116b1d-kube-api-access-zw865\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076558 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17f16f2-5d8f-4178-a6b5-879444c7b262-config-data\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-public-tls-certs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.076901 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8e508c6-7afc-4ae0-84bb-48dece116b1d-logs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.081530 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-public-tls-certs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.082399 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17f16f2-5d8f-4178-a6b5-879444c7b262-config-data\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.082857 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.085271 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.090520 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8e508c6-7afc-4ae0-84bb-48dece116b1d-config-data\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.092971 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mldtj\" (UniqueName: \"kubernetes.io/projected/c17f16f2-5d8f-4178-a6b5-879444c7b262-kube-api-access-mldtj\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.093002 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17f16f2-5d8f-4178-a6b5-879444c7b262-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c17f16f2-5d8f-4178-a6b5-879444c7b262\") " pod="openstack/nova-scheduler-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.096084 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw865\" (UniqueName: \"kubernetes.io/projected/e8e508c6-7afc-4ae0-84bb-48dece116b1d-kube-api-access-zw865\") pod \"nova-api-0\" (UID: \"e8e508c6-7afc-4ae0-84bb-48dece116b1d\") " pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.129618 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.151556 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.626292 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 15:38:03 crc kubenswrapper[4945]: W1008 15:38:03.636011 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc17f16f2_5d8f_4178_a6b5_879444c7b262.slice/crio-b8d752834c96ff19694ac458b98edba9900861691c845bee6dfade321e2e5a44 WatchSource:0}: Error finding container b8d752834c96ff19694ac458b98edba9900861691c845bee6dfade321e2e5a44: Status 404 returned error can't find the container with id b8d752834c96ff19694ac458b98edba9900861691c845bee6dfade321e2e5a44 Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.712040 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 15:38:03 crc kubenswrapper[4945]: W1008 15:38:03.717961 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8e508c6_7afc_4ae0_84bb_48dece116b1d.slice/crio-bef1236b01c262ff1312ea3a712f62f1906b3f72e63cc90890507e14165e3f2e WatchSource:0}: Error finding container bef1236b01c262ff1312ea3a712f62f1906b3f72e63cc90890507e14165e3f2e: Status 404 returned error can't find the container with id bef1236b01c262ff1312ea3a712f62f1906b3f72e63cc90890507e14165e3f2e Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.726838 4945 generic.go:334] "Generic (PLEG): container finished" podID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerID="1fbb94a8d8f61a0b6cef4a5fdb687b9157cf516718378738b1a4703dec50900b" exitCode=0 Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.726880 4945 generic.go:334] "Generic (PLEG): container finished" podID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerID="4711b6a17f7f102d9385a94f060dcdb2664bb77fd86d1db33c0cec5823c0ee92" exitCode=143 Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.726928 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"505a5605-d46e-4de8-bf48-0d1b32fa3971","Type":"ContainerDied","Data":"1fbb94a8d8f61a0b6cef4a5fdb687b9157cf516718378738b1a4703dec50900b"} Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.727027 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"505a5605-d46e-4de8-bf48-0d1b32fa3971","Type":"ContainerDied","Data":"4711b6a17f7f102d9385a94f060dcdb2664bb77fd86d1db33c0cec5823c0ee92"} Oct 08 15:38:03 crc kubenswrapper[4945]: I1008 15:38:03.728676 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c17f16f2-5d8f-4178-a6b5-879444c7b262","Type":"ContainerStarted","Data":"b8d752834c96ff19694ac458b98edba9900861691c845bee6dfade321e2e5a44"} Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.013306 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.041146 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fed2c6e-b5c7-4128-9aef-a310632ca057" path="/var/lib/kubelet/pods/6fed2c6e-b5c7-4128-9aef-a310632ca057/volumes" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.042408 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3a9136d-3567-4523-97bb-56936b7f6ae4" path="/var/lib/kubelet/pods/b3a9136d-3567-4523-97bb-56936b7f6ae4/volumes" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.100924 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-combined-ca-bundle\") pod \"505a5605-d46e-4de8-bf48-0d1b32fa3971\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.101009 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg9d6\" (UniqueName: \"kubernetes.io/projected/505a5605-d46e-4de8-bf48-0d1b32fa3971-kube-api-access-cg9d6\") pod \"505a5605-d46e-4de8-bf48-0d1b32fa3971\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.101043 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a5605-d46e-4de8-bf48-0d1b32fa3971-logs\") pod \"505a5605-d46e-4de8-bf48-0d1b32fa3971\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.101222 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-config-data\") pod \"505a5605-d46e-4de8-bf48-0d1b32fa3971\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.101261 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-nova-metadata-tls-certs\") pod \"505a5605-d46e-4de8-bf48-0d1b32fa3971\" (UID: \"505a5605-d46e-4de8-bf48-0d1b32fa3971\") " Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.103003 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/505a5605-d46e-4de8-bf48-0d1b32fa3971-logs" (OuterVolumeSpecName: "logs") pod "505a5605-d46e-4de8-bf48-0d1b32fa3971" (UID: "505a5605-d46e-4de8-bf48-0d1b32fa3971"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.108541 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/505a5605-d46e-4de8-bf48-0d1b32fa3971-kube-api-access-cg9d6" (OuterVolumeSpecName: "kube-api-access-cg9d6") pod "505a5605-d46e-4de8-bf48-0d1b32fa3971" (UID: "505a5605-d46e-4de8-bf48-0d1b32fa3971"). InnerVolumeSpecName "kube-api-access-cg9d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.171314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-config-data" (OuterVolumeSpecName: "config-data") pod "505a5605-d46e-4de8-bf48-0d1b32fa3971" (UID: "505a5605-d46e-4de8-bf48-0d1b32fa3971"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.172382 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "505a5605-d46e-4de8-bf48-0d1b32fa3971" (UID: "505a5605-d46e-4de8-bf48-0d1b32fa3971"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.189257 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "505a5605-d46e-4de8-bf48-0d1b32fa3971" (UID: "505a5605-d46e-4de8-bf48-0d1b32fa3971"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.204404 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg9d6\" (UniqueName: \"kubernetes.io/projected/505a5605-d46e-4de8-bf48-0d1b32fa3971-kube-api-access-cg9d6\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.204448 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a5605-d46e-4de8-bf48-0d1b32fa3971-logs\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.204462 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.204474 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.204487 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a5605-d46e-4de8-bf48-0d1b32fa3971-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.745424 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8e508c6-7afc-4ae0-84bb-48dece116b1d","Type":"ContainerStarted","Data":"93f1c2eaca6b89b32acf54840cc78371166c53e855e5de705c33e90a27806e5c"} Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.745467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8e508c6-7afc-4ae0-84bb-48dece116b1d","Type":"ContainerStarted","Data":"a74145c465df1827d70b777d4c5a42e9052bce520ce67709b067a4a0fce485c4"} Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.745478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e8e508c6-7afc-4ae0-84bb-48dece116b1d","Type":"ContainerStarted","Data":"bef1236b01c262ff1312ea3a712f62f1906b3f72e63cc90890507e14165e3f2e"} Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.752008 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"505a5605-d46e-4de8-bf48-0d1b32fa3971","Type":"ContainerDied","Data":"a868acf6cec282bbfb6b471bd12d132fa6d1c5a2bca743e8e8cd16b680011d3b"} Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.752045 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.752083 4945 scope.go:117] "RemoveContainer" containerID="1fbb94a8d8f61a0b6cef4a5fdb687b9157cf516718378738b1a4703dec50900b" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.767309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c17f16f2-5d8f-4178-a6b5-879444c7b262","Type":"ContainerStarted","Data":"025b8826be5da560e60e8d3450a5aebc38e2f0d278247257b35d9071a0215702"} Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.780433 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.780413564 podStartE2EDuration="2.780413564s" podCreationTimestamp="2025-10-08 15:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:38:04.774440838 +0000 UTC m=+1494.128355749" watchObservedRunningTime="2025-10-08 15:38:04.780413564 +0000 UTC m=+1494.134328465" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.806438 4945 scope.go:117] "RemoveContainer" containerID="4711b6a17f7f102d9385a94f060dcdb2664bb77fd86d1db33c0cec5823c0ee92" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.812650 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.812626621 podStartE2EDuration="2.812626621s" podCreationTimestamp="2025-10-08 15:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:38:04.799820248 +0000 UTC m=+1494.153735149" watchObservedRunningTime="2025-10-08 15:38:04.812626621 +0000 UTC m=+1494.166541532" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.832102 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.853320 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.868220 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:38:04 crc kubenswrapper[4945]: E1008 15:38:04.868815 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-log" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.868839 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-log" Oct 08 15:38:04 crc kubenswrapper[4945]: E1008 15:38:04.868861 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-metadata" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.868870 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-metadata" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.869185 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-metadata" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.869213 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" containerName="nova-metadata-log" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.870643 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.878500 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.881587 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 15:38:04 crc kubenswrapper[4945]: I1008 15:38:04.881769 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.022040 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-config-data\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.022169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzcl5\" (UniqueName: \"kubernetes.io/projected/1ce5750e-a063-42a1-bde2-0731599fa3bd-kube-api-access-hzcl5\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.022259 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.022450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ce5750e-a063-42a1-bde2-0731599fa3bd-logs\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.022894 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.124448 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-config-data\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.124767 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzcl5\" (UniqueName: \"kubernetes.io/projected/1ce5750e-a063-42a1-bde2-0731599fa3bd-kube-api-access-hzcl5\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.124962 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.125103 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ce5750e-a063-42a1-bde2-0731599fa3bd-logs\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.125267 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.125921 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ce5750e-a063-42a1-bde2-0731599fa3bd-logs\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.128054 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.128556 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.136301 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ce5750e-a063-42a1-bde2-0731599fa3bd-config-data\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.150054 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzcl5\" (UniqueName: \"kubernetes.io/projected/1ce5750e-a063-42a1-bde2-0731599fa3bd-kube-api-access-hzcl5\") pod \"nova-metadata-0\" (UID: \"1ce5750e-a063-42a1-bde2-0731599fa3bd\") " pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.199816 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.650536 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 15:38:05 crc kubenswrapper[4945]: W1008 15:38:05.654370 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ce5750e_a063_42a1_bde2_0731599fa3bd.slice/crio-a1fa774448298771911ef5b2e92bdbd07710eb622f441c7980b436736105423b WatchSource:0}: Error finding container a1fa774448298771911ef5b2e92bdbd07710eb622f441c7980b436736105423b: Status 404 returned error can't find the container with id a1fa774448298771911ef5b2e92bdbd07710eb622f441c7980b436736105423b Oct 08 15:38:05 crc kubenswrapper[4945]: I1008 15:38:05.778186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ce5750e-a063-42a1-bde2-0731599fa3bd","Type":"ContainerStarted","Data":"a1fa774448298771911ef5b2e92bdbd07710eb622f441c7980b436736105423b"} Oct 08 15:38:06 crc kubenswrapper[4945]: I1008 15:38:06.035790 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="505a5605-d46e-4de8-bf48-0d1b32fa3971" path="/var/lib/kubelet/pods/505a5605-d46e-4de8-bf48-0d1b32fa3971/volumes" Oct 08 15:38:06 crc kubenswrapper[4945]: I1008 15:38:06.796766 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ce5750e-a063-42a1-bde2-0731599fa3bd","Type":"ContainerStarted","Data":"3229df2c16841822f54a5f0bd6676d1f38cfa42ef6e11552695b7712a095bb9e"} Oct 08 15:38:06 crc kubenswrapper[4945]: I1008 15:38:06.797065 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1ce5750e-a063-42a1-bde2-0731599fa3bd","Type":"ContainerStarted","Data":"92dbdc4b89c6a5edf052673e3c25e1f9717f9db1b2345478a9e44b3156748be1"} Oct 08 15:38:06 crc kubenswrapper[4945]: I1008 15:38:06.823816 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.823798534 podStartE2EDuration="2.823798534s" podCreationTimestamp="2025-10-08 15:38:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:38:06.81581943 +0000 UTC m=+1496.169734331" watchObservedRunningTime="2025-10-08 15:38:06.823798534 +0000 UTC m=+1496.177713435" Oct 08 15:38:08 crc kubenswrapper[4945]: I1008 15:38:08.129784 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 15:38:10 crc kubenswrapper[4945]: I1008 15:38:10.200746 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 15:38:10 crc kubenswrapper[4945]: I1008 15:38:10.201089 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 15:38:13 crc kubenswrapper[4945]: I1008 15:38:13.130381 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 15:38:13 crc kubenswrapper[4945]: I1008 15:38:13.152398 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:38:13 crc kubenswrapper[4945]: I1008 15:38:13.152454 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 15:38:13 crc kubenswrapper[4945]: I1008 15:38:13.175518 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 15:38:13 crc kubenswrapper[4945]: I1008 15:38:13.936978 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 15:38:14 crc kubenswrapper[4945]: I1008 15:38:14.168260 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e8e508c6-7afc-4ae0-84bb-48dece116b1d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.223:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:38:14 crc kubenswrapper[4945]: I1008 15:38:14.168267 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e8e508c6-7afc-4ae0-84bb-48dece116b1d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.223:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:38:15 crc kubenswrapper[4945]: I1008 15:38:15.201318 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 15:38:15 crc kubenswrapper[4945]: I1008 15:38:15.201379 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 15:38:16 crc kubenswrapper[4945]: I1008 15:38:16.220435 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1ce5750e-a063-42a1-bde2-0731599fa3bd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.224:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:38:16 crc kubenswrapper[4945]: I1008 15:38:16.220450 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1ce5750e-a063-42a1-bde2-0731599fa3bd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.224:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.431039 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vzpsz"] Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.433681 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.455168 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vzpsz"] Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.489543 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-utilities\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.489607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-catalog-content\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.489823 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct99l\" (UniqueName: \"kubernetes.io/projected/84f40741-d080-4605-b0cf-35cbc0653a33-kube-api-access-ct99l\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.591576 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-utilities\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.591986 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-catalog-content\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.592021 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-utilities\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.592052 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct99l\" (UniqueName: \"kubernetes.io/projected/84f40741-d080-4605-b0cf-35cbc0653a33-kube-api-access-ct99l\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.592282 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-catalog-content\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.616738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct99l\" (UniqueName: \"kubernetes.io/projected/84f40741-d080-4605-b0cf-35cbc0653a33-kube-api-access-ct99l\") pod \"community-operators-vzpsz\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:18 crc kubenswrapper[4945]: I1008 15:38:18.767832 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:19 crc kubenswrapper[4945]: I1008 15:38:19.244430 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vzpsz"] Oct 08 15:38:19 crc kubenswrapper[4945]: I1008 15:38:19.926898 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 15:38:19 crc kubenswrapper[4945]: I1008 15:38:19.980331 4945 generic.go:334] "Generic (PLEG): container finished" podID="84f40741-d080-4605-b0cf-35cbc0653a33" containerID="00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5" exitCode=0 Oct 08 15:38:19 crc kubenswrapper[4945]: I1008 15:38:19.980397 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpsz" event={"ID":"84f40741-d080-4605-b0cf-35cbc0653a33","Type":"ContainerDied","Data":"00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5"} Oct 08 15:38:19 crc kubenswrapper[4945]: I1008 15:38:19.980435 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpsz" event={"ID":"84f40741-d080-4605-b0cf-35cbc0653a33","Type":"ContainerStarted","Data":"1bb8b15825eef9d552951b018f48e8a52b887687c3580c9d4f3c4928038e69db"} Oct 08 15:38:22 crc kubenswrapper[4945]: I1008 15:38:22.003906 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpsz" event={"ID":"84f40741-d080-4605-b0cf-35cbc0653a33","Type":"ContainerStarted","Data":"a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142"} Oct 08 15:38:23 crc kubenswrapper[4945]: I1008 15:38:23.014834 4945 generic.go:334] "Generic (PLEG): container finished" podID="84f40741-d080-4605-b0cf-35cbc0653a33" containerID="a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142" exitCode=0 Oct 08 15:38:23 crc kubenswrapper[4945]: I1008 15:38:23.014883 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpsz" event={"ID":"84f40741-d080-4605-b0cf-35cbc0653a33","Type":"ContainerDied","Data":"a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142"} Oct 08 15:38:23 crc kubenswrapper[4945]: I1008 15:38:23.164131 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 15:38:23 crc kubenswrapper[4945]: I1008 15:38:23.164753 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 15:38:23 crc kubenswrapper[4945]: I1008 15:38:23.166103 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 15:38:23 crc kubenswrapper[4945]: I1008 15:38:23.174143 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 15:38:24 crc kubenswrapper[4945]: I1008 15:38:24.040454 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 15:38:24 crc kubenswrapper[4945]: I1008 15:38:24.040789 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 15:38:24 crc kubenswrapper[4945]: I1008 15:38:24.040804 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpsz" event={"ID":"84f40741-d080-4605-b0cf-35cbc0653a33","Type":"ContainerStarted","Data":"f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87"} Oct 08 15:38:24 crc kubenswrapper[4945]: I1008 15:38:24.052416 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vzpsz" podStartSLOduration=2.4383127 podStartE2EDuration="6.052399604s" podCreationTimestamp="2025-10-08 15:38:18 +0000 UTC" firstStartedPulling="2025-10-08 15:38:19.983715625 +0000 UTC m=+1509.337630536" lastFinishedPulling="2025-10-08 15:38:23.597802539 +0000 UTC m=+1512.951717440" observedRunningTime="2025-10-08 15:38:24.046164673 +0000 UTC m=+1513.400079594" watchObservedRunningTime="2025-10-08 15:38:24.052399604 +0000 UTC m=+1513.406314505" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.209385 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.209477 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.217408 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.218328 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.823159 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b7dl7"] Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.825912 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.849669 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7dl7"] Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.930748 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-utilities\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.930850 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-catalog-content\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:25 crc kubenswrapper[4945]: I1008 15:38:25.930934 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcnd6\" (UniqueName: \"kubernetes.io/projected/4b0bde02-c61d-4d77-9ec7-77f719a354f1-kube-api-access-hcnd6\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.031968 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-catalog-content\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.032089 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcnd6\" (UniqueName: \"kubernetes.io/projected/4b0bde02-c61d-4d77-9ec7-77f719a354f1-kube-api-access-hcnd6\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.032234 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-utilities\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.032611 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-catalog-content\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.032779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-utilities\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.055637 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcnd6\" (UniqueName: \"kubernetes.io/projected/4b0bde02-c61d-4d77-9ec7-77f719a354f1-kube-api-access-hcnd6\") pod \"redhat-marketplace-b7dl7\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.150000 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:26 crc kubenswrapper[4945]: I1008 15:38:26.628902 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7dl7"] Oct 08 15:38:27 crc kubenswrapper[4945]: I1008 15:38:27.064061 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7dl7" event={"ID":"4b0bde02-c61d-4d77-9ec7-77f719a354f1","Type":"ContainerStarted","Data":"ed56f60700768b0717d6535e7cd8b0bb82fe239e382b5eb55adab595f8a08fe6"} Oct 08 15:38:28 crc kubenswrapper[4945]: I1008 15:38:28.078181 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerID="8289e3f9eeddb02f2374bb282ac70d3e256cb55e78a8a8af423d4533a687948c" exitCode=0 Oct 08 15:38:28 crc kubenswrapper[4945]: I1008 15:38:28.078297 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7dl7" event={"ID":"4b0bde02-c61d-4d77-9ec7-77f719a354f1","Type":"ContainerDied","Data":"8289e3f9eeddb02f2374bb282ac70d3e256cb55e78a8a8af423d4533a687948c"} Oct 08 15:38:28 crc kubenswrapper[4945]: I1008 15:38:28.768076 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:28 crc kubenswrapper[4945]: I1008 15:38:28.768530 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:28 crc kubenswrapper[4945]: I1008 15:38:28.866394 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:29 crc kubenswrapper[4945]: I1008 15:38:29.151002 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:30 crc kubenswrapper[4945]: I1008 15:38:30.102883 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerID="7cd84effde7b665808233b56e0a46d433135176c58a1ca21c3db2500847bb495" exitCode=0 Oct 08 15:38:30 crc kubenswrapper[4945]: I1008 15:38:30.103066 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7dl7" event={"ID":"4b0bde02-c61d-4d77-9ec7-77f719a354f1","Type":"ContainerDied","Data":"7cd84effde7b665808233b56e0a46d433135176c58a1ca21c3db2500847bb495"} Oct 08 15:38:30 crc kubenswrapper[4945]: I1008 15:38:30.809013 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vzpsz"] Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.118338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7dl7" event={"ID":"4b0bde02-c61d-4d77-9ec7-77f719a354f1","Type":"ContainerStarted","Data":"7944122f426088debc84d17e6bcf7fcb196e3fec9651fab7be86ccee136061e8"} Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.118486 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vzpsz" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="registry-server" containerID="cri-o://f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87" gracePeriod=2 Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.145654 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b7dl7" podStartSLOduration=3.6189113859999997 podStartE2EDuration="6.145636045s" podCreationTimestamp="2025-10-08 15:38:25 +0000 UTC" firstStartedPulling="2025-10-08 15:38:28.083090896 +0000 UTC m=+1517.437005807" lastFinishedPulling="2025-10-08 15:38:30.609815555 +0000 UTC m=+1519.963730466" observedRunningTime="2025-10-08 15:38:31.134516573 +0000 UTC m=+1520.488431474" watchObservedRunningTime="2025-10-08 15:38:31.145636045 +0000 UTC m=+1520.499550946" Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.570300 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.642405 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-utilities\") pod \"84f40741-d080-4605-b0cf-35cbc0653a33\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.642473 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct99l\" (UniqueName: \"kubernetes.io/projected/84f40741-d080-4605-b0cf-35cbc0653a33-kube-api-access-ct99l\") pod \"84f40741-d080-4605-b0cf-35cbc0653a33\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.642704 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-catalog-content\") pod \"84f40741-d080-4605-b0cf-35cbc0653a33\" (UID: \"84f40741-d080-4605-b0cf-35cbc0653a33\") " Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.643231 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-utilities" (OuterVolumeSpecName: "utilities") pod "84f40741-d080-4605-b0cf-35cbc0653a33" (UID: "84f40741-d080-4605-b0cf-35cbc0653a33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.650319 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84f40741-d080-4605-b0cf-35cbc0653a33-kube-api-access-ct99l" (OuterVolumeSpecName: "kube-api-access-ct99l") pod "84f40741-d080-4605-b0cf-35cbc0653a33" (UID: "84f40741-d080-4605-b0cf-35cbc0653a33"). InnerVolumeSpecName "kube-api-access-ct99l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.690306 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84f40741-d080-4605-b0cf-35cbc0653a33" (UID: "84f40741-d080-4605-b0cf-35cbc0653a33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.744664 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.744705 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f40741-d080-4605-b0cf-35cbc0653a33-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:31 crc kubenswrapper[4945]: I1008 15:38:31.744717 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct99l\" (UniqueName: \"kubernetes.io/projected/84f40741-d080-4605-b0cf-35cbc0653a33-kube-api-access-ct99l\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.130858 4945 generic.go:334] "Generic (PLEG): container finished" podID="84f40741-d080-4605-b0cf-35cbc0653a33" containerID="f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87" exitCode=0 Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.130899 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpsz" event={"ID":"84f40741-d080-4605-b0cf-35cbc0653a33","Type":"ContainerDied","Data":"f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87"} Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.130909 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpsz" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.130933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpsz" event={"ID":"84f40741-d080-4605-b0cf-35cbc0653a33","Type":"ContainerDied","Data":"1bb8b15825eef9d552951b018f48e8a52b887687c3580c9d4f3c4928038e69db"} Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.130954 4945 scope.go:117] "RemoveContainer" containerID="f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.159757 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vzpsz"] Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.162954 4945 scope.go:117] "RemoveContainer" containerID="a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.172886 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vzpsz"] Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.184670 4945 scope.go:117] "RemoveContainer" containerID="00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.253306 4945 scope.go:117] "RemoveContainer" containerID="f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87" Oct 08 15:38:32 crc kubenswrapper[4945]: E1008 15:38:32.254550 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87\": container with ID starting with f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87 not found: ID does not exist" containerID="f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.254596 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87"} err="failed to get container status \"f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87\": rpc error: code = NotFound desc = could not find container \"f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87\": container with ID starting with f7b7c145d6f037430754a9fef52fa2bd760e7c16479b22c9d7bb37303213fd87 not found: ID does not exist" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.254633 4945 scope.go:117] "RemoveContainer" containerID="a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142" Oct 08 15:38:32 crc kubenswrapper[4945]: E1008 15:38:32.255151 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142\": container with ID starting with a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142 not found: ID does not exist" containerID="a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.255194 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142"} err="failed to get container status \"a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142\": rpc error: code = NotFound desc = could not find container \"a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142\": container with ID starting with a16ff105c61ad9f50bc17183bd2e0f0ff770ec950f473d33526ffd96a75c1142 not found: ID does not exist" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.255227 4945 scope.go:117] "RemoveContainer" containerID="00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5" Oct 08 15:38:32 crc kubenswrapper[4945]: E1008 15:38:32.255639 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5\": container with ID starting with 00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5 not found: ID does not exist" containerID="00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5" Oct 08 15:38:32 crc kubenswrapper[4945]: I1008 15:38:32.255740 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5"} err="failed to get container status \"00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5\": rpc error: code = NotFound desc = could not find container \"00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5\": container with ID starting with 00def1bdd42a33c9f741f81013367186b5030504a15bffdea711a17811d4d8e5 not found: ID does not exist" Oct 08 15:38:34 crc kubenswrapper[4945]: I1008 15:38:34.041072 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" path="/var/lib/kubelet/pods/84f40741-d080-4605-b0cf-35cbc0653a33/volumes" Oct 08 15:38:34 crc kubenswrapper[4945]: I1008 15:38:34.064879 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:38:34 crc kubenswrapper[4945]: I1008 15:38:34.902161 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:38:36 crc kubenswrapper[4945]: I1008 15:38:36.150447 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:36 crc kubenswrapper[4945]: I1008 15:38:36.152666 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:36 crc kubenswrapper[4945]: I1008 15:38:36.211226 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:37 crc kubenswrapper[4945]: I1008 15:38:37.228368 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:37 crc kubenswrapper[4945]: I1008 15:38:37.278778 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7dl7"] Oct 08 15:38:37 crc kubenswrapper[4945]: I1008 15:38:37.503146 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="rabbitmq" containerID="cri-o://8a07f234ae8ef9ef30b98aaf68ce0a8496f264fed9c6a782c9995e5b987b152b" gracePeriod=604797 Oct 08 15:38:38 crc kubenswrapper[4945]: I1008 15:38:38.293827 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="rabbitmq" containerID="cri-o://4e408d857f77f1cce0c74bb1e177aeff00c716bca702c1487d3a63e4cdf3fba0" gracePeriod=604797 Oct 08 15:38:39 crc kubenswrapper[4945]: I1008 15:38:39.206275 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec56f6d2-9c69-47f1-b834-169de890c512" containerID="8a07f234ae8ef9ef30b98aaf68ce0a8496f264fed9c6a782c9995e5b987b152b" exitCode=0 Oct 08 15:38:39 crc kubenswrapper[4945]: I1008 15:38:39.206395 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec56f6d2-9c69-47f1-b834-169de890c512","Type":"ContainerDied","Data":"8a07f234ae8ef9ef30b98aaf68ce0a8496f264fed9c6a782c9995e5b987b152b"} Oct 08 15:38:39 crc kubenswrapper[4945]: I1008 15:38:39.206965 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b7dl7" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="registry-server" containerID="cri-o://7944122f426088debc84d17e6bcf7fcb196e3fec9651fab7be86ccee136061e8" gracePeriod=2 Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.231714 4945 generic.go:334] "Generic (PLEG): container finished" podID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerID="4e408d857f77f1cce0c74bb1e177aeff00c716bca702c1487d3a63e4cdf3fba0" exitCode=0 Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.231986 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2481ce07-e2b2-4afb-b5f1-027bfefd0f38","Type":"ContainerDied","Data":"4e408d857f77f1cce0c74bb1e177aeff00c716bca702c1487d3a63e4cdf3fba0"} Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.234671 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerID="7944122f426088debc84d17e6bcf7fcb196e3fec9651fab7be86ccee136061e8" exitCode=0 Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.234692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7dl7" event={"ID":"4b0bde02-c61d-4d77-9ec7-77f719a354f1","Type":"ContainerDied","Data":"7944122f426088debc84d17e6bcf7fcb196e3fec9651fab7be86ccee136061e8"} Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.234707 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7dl7" event={"ID":"4b0bde02-c61d-4d77-9ec7-77f719a354f1","Type":"ContainerDied","Data":"ed56f60700768b0717d6535e7cd8b0bb82fe239e382b5eb55adab595f8a08fe6"} Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.234717 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed56f60700768b0717d6535e7cd8b0bb82fe239e382b5eb55adab595f8a08fe6" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.317363 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.325439 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460060 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j96z6\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-kube-api-access-j96z6\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460114 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-server-conf\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-utilities\") pod \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460317 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-catalog-content\") pod \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460474 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-plugins-conf\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460544 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-config-data\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460579 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcnd6\" (UniqueName: \"kubernetes.io/projected/4b0bde02-c61d-4d77-9ec7-77f719a354f1-kube-api-access-hcnd6\") pod \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\" (UID: \"4b0bde02-c61d-4d77-9ec7-77f719a354f1\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460656 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-tls\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460745 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec56f6d2-9c69-47f1-b834-169de890c512-pod-info\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460785 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec56f6d2-9c69-47f1-b834-169de890c512-erlang-cookie-secret\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460815 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-erlang-cookie\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460844 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-plugins\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.460890 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-confd\") pod \"ec56f6d2-9c69-47f1-b834-169de890c512\" (UID: \"ec56f6d2-9c69-47f1-b834-169de890c512\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.465794 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-utilities" (OuterVolumeSpecName: "utilities") pod "4b0bde02-c61d-4d77-9ec7-77f719a354f1" (UID: "4b0bde02-c61d-4d77-9ec7-77f719a354f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.484520 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b0bde02-c61d-4d77-9ec7-77f719a354f1" (UID: "4b0bde02-c61d-4d77-9ec7-77f719a354f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.490928 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b0bde02-c61d-4d77-9ec7-77f719a354f1-kube-api-access-hcnd6" (OuterVolumeSpecName: "kube-api-access-hcnd6") pod "4b0bde02-c61d-4d77-9ec7-77f719a354f1" (UID: "4b0bde02-c61d-4d77-9ec7-77f719a354f1"). InnerVolumeSpecName "kube-api-access-hcnd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.526446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.563318 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.563345 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b0bde02-c61d-4d77-9ec7-77f719a354f1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.563369 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.563378 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcnd6\" (UniqueName: \"kubernetes.io/projected/4b0bde02-c61d-4d77-9ec7-77f719a354f1-kube-api-access-hcnd6\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.597014 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.632210 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ec56f6d2-9c69-47f1-b834-169de890c512-pod-info" (OuterVolumeSpecName: "pod-info") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.632241 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec56f6d2-9c69-47f1-b834-169de890c512-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.635212 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-kube-api-access-j96z6" (OuterVolumeSpecName: "kube-api-access-j96z6") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "kube-api-access-j96z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.643065 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.643216 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.643221 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-config-data" (OuterVolumeSpecName: "config-data") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.647401 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.647425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.664851 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-server-conf" (OuterVolumeSpecName: "server-conf") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667256 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667309 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j96z6\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-kube-api-access-j96z6\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667321 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667330 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667339 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec56f6d2-9c69-47f1-b834-169de890c512-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667347 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667356 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec56f6d2-9c69-47f1-b834-169de890c512-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667366 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec56f6d2-9c69-47f1-b834-169de890c512-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667377 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.667389 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.686077 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ec56f6d2-9c69-47f1-b834-169de890c512" (UID: "ec56f6d2-9c69-47f1-b834-169de890c512"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.771907 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec56f6d2-9c69-47f1-b834-169de890c512-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.870008 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.976977 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-confd\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977057 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-config-data\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977182 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-tls\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977199 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977244 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-server-conf\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977296 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-plugins-conf\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977316 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-pod-info\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977339 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-plugins\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-erlang-cookie-secret\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977421 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctr6t\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-kube-api-access-ctr6t\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.977472 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-erlang-cookie\") pod \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\" (UID: \"2481ce07-e2b2-4afb-b5f1-027bfefd0f38\") " Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.981055 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.982347 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.983597 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.985541 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.987214 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-pod-info" (OuterVolumeSpecName: "pod-info") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.989328 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.991519 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-kube-api-access-ctr6t" (OuterVolumeSpecName: "kube-api-access-ctr6t") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "kube-api-access-ctr6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:40 crc kubenswrapper[4945]: I1008 15:38:40.996227 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.025686 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-config-data" (OuterVolumeSpecName: "config-data") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.056603 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-server-conf" (OuterVolumeSpecName: "server-conf") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080491 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080528 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080538 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080546 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080555 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080564 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctr6t\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-kube-api-access-ctr6t\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080572 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080581 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080589 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.080624 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.100202 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.123197 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2481ce07-e2b2-4afb-b5f1-027bfefd0f38" (UID: "2481ce07-e2b2-4afb-b5f1-027bfefd0f38"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.181823 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2481ce07-e2b2-4afb-b5f1-027bfefd0f38-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.182219 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.247666 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec56f6d2-9c69-47f1-b834-169de890c512","Type":"ContainerDied","Data":"d0b11f27eb8a2cda5aa8447d50ed22c0ff12f08930346537aa8d7133cd1128a9"} Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.247717 4945 scope.go:117] "RemoveContainer" containerID="8a07f234ae8ef9ef30b98aaf68ce0a8496f264fed9c6a782c9995e5b987b152b" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.247733 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.251415 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7dl7" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.252265 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2481ce07-e2b2-4afb-b5f1-027bfefd0f38","Type":"ContainerDied","Data":"01b5efbe7b0c8a7e069068266e7d5fc93815f4a6d6b4123ce36a261dd2ca6342"} Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.252478 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.297997 4945 scope.go:117] "RemoveContainer" containerID="de13a17be1ddba6e92ab07e7c7fda48d8d412d56c677f0cfff1aa52301f80782" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.356174 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.381277 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.414293 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.419381 4945 scope.go:117] "RemoveContainer" containerID="4e408d857f77f1cce0c74bb1e177aeff00c716bca702c1487d3a63e4cdf3fba0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.428549 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.428988 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="extract-utilities" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429006 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="extract-utilities" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429023 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="extract-utilities" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429029 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="extract-utilities" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429038 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="registry-server" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429043 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="registry-server" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429054 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="setup-container" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429060 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="setup-container" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429070 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="rabbitmq" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429078 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="rabbitmq" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429093 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="extract-content" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429099 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="extract-content" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429135 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="setup-container" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429141 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="setup-container" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429152 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="extract-content" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429158 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="extract-content" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429169 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="registry-server" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429174 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="registry-server" Oct 08 15:38:41 crc kubenswrapper[4945]: E1008 15:38:41.429184 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="rabbitmq" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429190 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="rabbitmq" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429349 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" containerName="registry-server" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429366 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="84f40741-d080-4605-b0cf-35cbc0653a33" containerName="registry-server" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429385 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" containerName="rabbitmq" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.429401 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" containerName="rabbitmq" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.430473 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.436278 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.436330 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.436278 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.436444 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.436491 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.436620 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.436753 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mqvcn" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.440345 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.465636 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.465785 4945 scope.go:117] "RemoveContainer" containerID="7e947c42db08811c5f07c0d5c81d80376a93939789814059a11b707a33d120e8" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.473106 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.475544 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.478187 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.478377 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lcgsm" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.481454 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.481673 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.481959 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.482229 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.485418 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.504983 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7dl7"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.526072 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7dl7"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.535239 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592458 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592528 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e51e86c4-2107-417e-9a3d-383d7a60da48-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592566 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592590 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592623 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592645 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e51e86c4-2107-417e-9a3d-383d7a60da48-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592665 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592793 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592876 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-config-data\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592894 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592916 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592936 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.592968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593007 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593022 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9hfc\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-kube-api-access-m9hfc\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593049 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593075 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593102 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593144 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx9xg\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-kube-api-access-xx9xg\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593179 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593209 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.593225 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695240 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695289 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e51e86c4-2107-417e-9a3d-383d7a60da48-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695346 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695408 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e51e86c4-2107-417e-9a3d-383d7a60da48-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695432 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695454 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695484 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-config-data\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695498 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.696370 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.696595 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.696903 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.696944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.695504 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697010 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697038 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697082 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697188 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697213 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9hfc\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-kube-api-access-m9hfc\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697256 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697298 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697344 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697371 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx9xg\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-kube-api-access-xx9xg\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697411 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697451 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697462 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697479 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.697656 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-config-data\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.698470 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e51e86c4-2107-417e-9a3d-383d7a60da48-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.699854 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.700059 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.700318 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.700479 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.700788 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.700806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e51e86c4-2107-417e-9a3d-383d7a60da48-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.703342 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.709061 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.713186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.716190 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.716841 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.716956 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e51e86c4-2107-417e-9a3d-383d7a60da48-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.719155 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9hfc\" (UniqueName: \"kubernetes.io/projected/e51e86c4-2107-417e-9a3d-383d7a60da48-kube-api-access-m9hfc\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.719825 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx9xg\" (UniqueName: \"kubernetes.io/projected/a252dc40-e13b-400d-a2b4-8b7e99d9eca1-kube-api-access-xx9xg\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.747186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a252dc40-e13b-400d-a2b4-8b7e99d9eca1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.764134 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.773299 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e51e86c4-2107-417e-9a3d-383d7a60da48\") " pod="openstack/rabbitmq-server-0" Oct 08 15:38:41 crc kubenswrapper[4945]: I1008 15:38:41.833897 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 15:38:42 crc kubenswrapper[4945]: I1008 15:38:42.060017 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2481ce07-e2b2-4afb-b5f1-027bfefd0f38" path="/var/lib/kubelet/pods/2481ce07-e2b2-4afb-b5f1-027bfefd0f38/volumes" Oct 08 15:38:42 crc kubenswrapper[4945]: I1008 15:38:42.073510 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b0bde02-c61d-4d77-9ec7-77f719a354f1" path="/var/lib/kubelet/pods/4b0bde02-c61d-4d77-9ec7-77f719a354f1/volumes" Oct 08 15:38:42 crc kubenswrapper[4945]: I1008 15:38:42.074891 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec56f6d2-9c69-47f1-b834-169de890c512" path="/var/lib/kubelet/pods/ec56f6d2-9c69-47f1-b834-169de890c512/volumes" Oct 08 15:38:42 crc kubenswrapper[4945]: I1008 15:38:42.362421 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 15:38:42 crc kubenswrapper[4945]: I1008 15:38:42.539682 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 15:38:43 crc kubenswrapper[4945]: I1008 15:38:43.275943 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e51e86c4-2107-417e-9a3d-383d7a60da48","Type":"ContainerStarted","Data":"e0cd6951427b4e159de0708ec8d68857da9ffcd8622ef1593f68e6d947d28adf"} Oct 08 15:38:43 crc kubenswrapper[4945]: I1008 15:38:43.280110 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a252dc40-e13b-400d-a2b4-8b7e99d9eca1","Type":"ContainerStarted","Data":"4d3a4da6e315d50ae2d4d9c1f094f2ec20abda4bd737a003ca492cad932c64a6"} Oct 08 15:38:44 crc kubenswrapper[4945]: I1008 15:38:44.290424 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a252dc40-e13b-400d-a2b4-8b7e99d9eca1","Type":"ContainerStarted","Data":"b36f833544bf48a858c599b0ee013f7e3e8a075cde7a98be7ffec3c49972cc48"} Oct 08 15:38:44 crc kubenswrapper[4945]: I1008 15:38:44.291959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e51e86c4-2107-417e-9a3d-383d7a60da48","Type":"ContainerStarted","Data":"eb70d577a2b8a769b990acc4892ed74d29111d41d0e07f0fcb2ccea30037f24f"} Oct 08 15:38:49 crc kubenswrapper[4945]: I1008 15:38:49.184076 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:38:49 crc kubenswrapper[4945]: I1008 15:38:49.184784 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.051538 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b8889b6c9-m9r4w"] Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.054906 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.057737 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.076397 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b8889b6c9-m9r4w"] Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.191705 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-openstack-edpm-ipam\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.191777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-swift-storage-0\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.191796 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-svc\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.191952 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-config\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.192177 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-sb\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.192450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l6cf\" (UniqueName: \"kubernetes.io/projected/871f25c1-f5b7-4d76-94bd-e93a39fb3724-kube-api-access-5l6cf\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.192648 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-nb\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.294786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l6cf\" (UniqueName: \"kubernetes.io/projected/871f25c1-f5b7-4d76-94bd-e93a39fb3724-kube-api-access-5l6cf\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.294890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-nb\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.294966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-openstack-edpm-ipam\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.295001 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-swift-storage-0\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.295033 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-svc\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.295067 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-config\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.295126 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-sb\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.296015 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-config\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.296300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-swift-storage-0\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.296462 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-sb\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.296547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-nb\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.296669 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-svc\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.296674 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-openstack-edpm-ipam\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.320156 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l6cf\" (UniqueName: \"kubernetes.io/projected/871f25c1-f5b7-4d76-94bd-e93a39fb3724-kube-api-access-5l6cf\") pod \"dnsmasq-dns-7b8889b6c9-m9r4w\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.388048 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:55 crc kubenswrapper[4945]: I1008 15:38:55.913050 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b8889b6c9-m9r4w"] Oct 08 15:38:56 crc kubenswrapper[4945]: I1008 15:38:56.419817 4945 generic.go:334] "Generic (PLEG): container finished" podID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerID="964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396" exitCode=0 Oct 08 15:38:56 crc kubenswrapper[4945]: I1008 15:38:56.419870 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" event={"ID":"871f25c1-f5b7-4d76-94bd-e93a39fb3724","Type":"ContainerDied","Data":"964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396"} Oct 08 15:38:56 crc kubenswrapper[4945]: I1008 15:38:56.420161 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" event={"ID":"871f25c1-f5b7-4d76-94bd-e93a39fb3724","Type":"ContainerStarted","Data":"a8798d6563df637eedce27d76b936d9ccf5e53dcb249d23809ead3be7627744f"} Oct 08 15:38:57 crc kubenswrapper[4945]: I1008 15:38:57.433822 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" event={"ID":"871f25c1-f5b7-4d76-94bd-e93a39fb3724","Type":"ContainerStarted","Data":"310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc"} Oct 08 15:38:57 crc kubenswrapper[4945]: I1008 15:38:57.436365 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:38:57 crc kubenswrapper[4945]: I1008 15:38:57.467382 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" podStartSLOduration=2.467353093 podStartE2EDuration="2.467353093s" podCreationTimestamp="2025-10-08 15:38:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:38:57.459960492 +0000 UTC m=+1546.813875403" watchObservedRunningTime="2025-10-08 15:38:57.467353093 +0000 UTC m=+1546.821268024" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.389263 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.489274 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766ff48cb5-86ctl"] Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.489918 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" podUID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerName="dnsmasq-dns" containerID="cri-o://a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d" gracePeriod=10 Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.603813 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f7b68cdfc-bhvjv"] Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.607644 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.623235 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f7b68cdfc-bhvjv"] Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.729567 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-ovsdbserver-nb\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.729625 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-config\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.729687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlgw9\" (UniqueName: \"kubernetes.io/projected/108d033e-a880-4b1d-ba80-293d36d0a58a-kube-api-access-rlgw9\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.729714 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-ovsdbserver-sb\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.729740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-dns-svc\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.729818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.729883 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-dns-swift-storage-0\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.832545 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-dns-swift-storage-0\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.832674 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-ovsdbserver-nb\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.832702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-config\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.832753 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlgw9\" (UniqueName: \"kubernetes.io/projected/108d033e-a880-4b1d-ba80-293d36d0a58a-kube-api-access-rlgw9\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.832779 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-ovsdbserver-sb\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.832806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-dns-svc\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.832881 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.833940 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.834125 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-config\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.834943 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-dns-swift-storage-0\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.835694 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-ovsdbserver-nb\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.836137 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-dns-svc\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.836309 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/108d033e-a880-4b1d-ba80-293d36d0a58a-ovsdbserver-sb\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.856161 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlgw9\" (UniqueName: \"kubernetes.io/projected/108d033e-a880-4b1d-ba80-293d36d0a58a-kube-api-access-rlgw9\") pod \"dnsmasq-dns-6f7b68cdfc-bhvjv\" (UID: \"108d033e-a880-4b1d-ba80-293d36d0a58a\") " pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:05 crc kubenswrapper[4945]: I1008 15:39:05.991980 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.105963 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.241321 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-nb\") pod \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.241505 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-config\") pod \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.241584 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-svc\") pod \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.241763 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-sb\") pod \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.241848 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-swift-storage-0\") pod \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.241906 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7ctk\" (UniqueName: \"kubernetes.io/projected/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-kube-api-access-p7ctk\") pod \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\" (UID: \"6811fe37-bd5a-4a52-91a6-91be2ad24e4c\") " Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.249759 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-kube-api-access-p7ctk" (OuterVolumeSpecName: "kube-api-access-p7ctk") pod "6811fe37-bd5a-4a52-91a6-91be2ad24e4c" (UID: "6811fe37-bd5a-4a52-91a6-91be2ad24e4c"). InnerVolumeSpecName "kube-api-access-p7ctk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.303057 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6811fe37-bd5a-4a52-91a6-91be2ad24e4c" (UID: "6811fe37-bd5a-4a52-91a6-91be2ad24e4c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.325003 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6811fe37-bd5a-4a52-91a6-91be2ad24e4c" (UID: "6811fe37-bd5a-4a52-91a6-91be2ad24e4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.341039 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-config" (OuterVolumeSpecName: "config") pod "6811fe37-bd5a-4a52-91a6-91be2ad24e4c" (UID: "6811fe37-bd5a-4a52-91a6-91be2ad24e4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.344433 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6811fe37-bd5a-4a52-91a6-91be2ad24e4c" (UID: "6811fe37-bd5a-4a52-91a6-91be2ad24e4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.344850 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.344879 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7ctk\" (UniqueName: \"kubernetes.io/projected/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-kube-api-access-p7ctk\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.344889 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.344900 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.344910 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.351213 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6811fe37-bd5a-4a52-91a6-91be2ad24e4c" (UID: "6811fe37-bd5a-4a52-91a6-91be2ad24e4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.444975 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f7b68cdfc-bhvjv"] Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.446502 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6811fe37-bd5a-4a52-91a6-91be2ad24e4c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:06 crc kubenswrapper[4945]: W1008 15:39:06.447628 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod108d033e_a880_4b1d_ba80_293d36d0a58a.slice/crio-8dc6a2be177b0cd83c964f31d6bf1a022df8fea949780c608834a49bbe88d5c5 WatchSource:0}: Error finding container 8dc6a2be177b0cd83c964f31d6bf1a022df8fea949780c608834a49bbe88d5c5: Status 404 returned error can't find the container with id 8dc6a2be177b0cd83c964f31d6bf1a022df8fea949780c608834a49bbe88d5c5 Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.528659 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" event={"ID":"108d033e-a880-4b1d-ba80-293d36d0a58a","Type":"ContainerStarted","Data":"8dc6a2be177b0cd83c964f31d6bf1a022df8fea949780c608834a49bbe88d5c5"} Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.530908 4945 generic.go:334] "Generic (PLEG): container finished" podID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerID="a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d" exitCode=0 Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.530947 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" event={"ID":"6811fe37-bd5a-4a52-91a6-91be2ad24e4c","Type":"ContainerDied","Data":"a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d"} Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.530955 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.530975 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766ff48cb5-86ctl" event={"ID":"6811fe37-bd5a-4a52-91a6-91be2ad24e4c","Type":"ContainerDied","Data":"2599d2af091cfdc1f035a278b3700769e94adda07526463d4073da6c96419fc8"} Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.530992 4945 scope.go:117] "RemoveContainer" containerID="a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.588034 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766ff48cb5-86ctl"] Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.590418 4945 scope.go:117] "RemoveContainer" containerID="30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.602602 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-766ff48cb5-86ctl"] Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.640549 4945 scope.go:117] "RemoveContainer" containerID="a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d" Oct 08 15:39:06 crc kubenswrapper[4945]: E1008 15:39:06.641126 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d\": container with ID starting with a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d not found: ID does not exist" containerID="a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.641170 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d"} err="failed to get container status \"a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d\": rpc error: code = NotFound desc = could not find container \"a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d\": container with ID starting with a4656194c7f72f205b69a46b678b2774c39844766997c077267154e4ad07314d not found: ID does not exist" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.641196 4945 scope.go:117] "RemoveContainer" containerID="30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c" Oct 08 15:39:06 crc kubenswrapper[4945]: E1008 15:39:06.641520 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c\": container with ID starting with 30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c not found: ID does not exist" containerID="30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c" Oct 08 15:39:06 crc kubenswrapper[4945]: I1008 15:39:06.641552 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c"} err="failed to get container status \"30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c\": rpc error: code = NotFound desc = could not find container \"30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c\": container with ID starting with 30ed0d8d8c06ade6ad643822a7fd99c504b715c405c3d227a9bc9c055261d02c not found: ID does not exist" Oct 08 15:39:07 crc kubenswrapper[4945]: I1008 15:39:07.544827 4945 generic.go:334] "Generic (PLEG): container finished" podID="108d033e-a880-4b1d-ba80-293d36d0a58a" containerID="10b321fc792bef8d1dbb8d153617a7ea0a3d2f3d17320313ee4016b3480a55e0" exitCode=0 Oct 08 15:39:07 crc kubenswrapper[4945]: I1008 15:39:07.544875 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" event={"ID":"108d033e-a880-4b1d-ba80-293d36d0a58a","Type":"ContainerDied","Data":"10b321fc792bef8d1dbb8d153617a7ea0a3d2f3d17320313ee4016b3480a55e0"} Oct 08 15:39:08 crc kubenswrapper[4945]: I1008 15:39:08.035420 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" path="/var/lib/kubelet/pods/6811fe37-bd5a-4a52-91a6-91be2ad24e4c/volumes" Oct 08 15:39:08 crc kubenswrapper[4945]: I1008 15:39:08.559301 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" event={"ID":"108d033e-a880-4b1d-ba80-293d36d0a58a","Type":"ContainerStarted","Data":"00c31e8ece523061d371d1a051a1ea7f7fc31149a0291abb8a6c522d26597f2b"} Oct 08 15:39:08 crc kubenswrapper[4945]: I1008 15:39:08.559606 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:08 crc kubenswrapper[4945]: I1008 15:39:08.596269 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" podStartSLOduration=3.5962463380000003 podStartE2EDuration="3.596246338s" podCreationTimestamp="2025-10-08 15:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:39:08.58571162 +0000 UTC m=+1557.939626551" watchObservedRunningTime="2025-10-08 15:39:08.596246338 +0000 UTC m=+1557.950161259" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.160060 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bmlfk"] Oct 08 15:39:13 crc kubenswrapper[4945]: E1008 15:39:13.161724 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerName="dnsmasq-dns" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.161755 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerName="dnsmasq-dns" Oct 08 15:39:13 crc kubenswrapper[4945]: E1008 15:39:13.161828 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerName="init" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.161837 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerName="init" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.162501 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6811fe37-bd5a-4a52-91a6-91be2ad24e4c" containerName="dnsmasq-dns" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.168400 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.195444 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bmlfk"] Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.298732 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9vh9\" (UniqueName: \"kubernetes.io/projected/271df054-2fac-48c5-aa11-898d24437cb3-kube-api-access-c9vh9\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.299287 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-utilities\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.299537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-catalog-content\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.401502 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9vh9\" (UniqueName: \"kubernetes.io/projected/271df054-2fac-48c5-aa11-898d24437cb3-kube-api-access-c9vh9\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.401641 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-utilities\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.401736 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-catalog-content\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.402413 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-catalog-content\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.402645 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-utilities\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.422390 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9vh9\" (UniqueName: \"kubernetes.io/projected/271df054-2fac-48c5-aa11-898d24437cb3-kube-api-access-c9vh9\") pod \"certified-operators-bmlfk\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:13 crc kubenswrapper[4945]: I1008 15:39:13.502235 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:14 crc kubenswrapper[4945]: I1008 15:39:14.019135 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bmlfk"] Oct 08 15:39:14 crc kubenswrapper[4945]: I1008 15:39:14.662348 4945 generic.go:334] "Generic (PLEG): container finished" podID="271df054-2fac-48c5-aa11-898d24437cb3" containerID="d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796" exitCode=0 Oct 08 15:39:14 crc kubenswrapper[4945]: I1008 15:39:14.662442 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmlfk" event={"ID":"271df054-2fac-48c5-aa11-898d24437cb3","Type":"ContainerDied","Data":"d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796"} Oct 08 15:39:14 crc kubenswrapper[4945]: I1008 15:39:14.662837 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmlfk" event={"ID":"271df054-2fac-48c5-aa11-898d24437cb3","Type":"ContainerStarted","Data":"8b59a35ecacd7b533700ecd0e92d59b36e81cb69199eaf489636f031633d0bb6"} Oct 08 15:39:15 crc kubenswrapper[4945]: I1008 15:39:15.993261 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f7b68cdfc-bhvjv" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.051340 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b8889b6c9-m9r4w"] Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.051612 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" podUID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerName="dnsmasq-dns" containerID="cri-o://310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc" gracePeriod=10 Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.686359 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmlfk" event={"ID":"271df054-2fac-48c5-aa11-898d24437cb3","Type":"ContainerStarted","Data":"dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94"} Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.686454 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.687786 4945 generic.go:334] "Generic (PLEG): container finished" podID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerID="310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc" exitCode=0 Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.687845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" event={"ID":"871f25c1-f5b7-4d76-94bd-e93a39fb3724","Type":"ContainerDied","Data":"310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc"} Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.687874 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" event={"ID":"871f25c1-f5b7-4d76-94bd-e93a39fb3724","Type":"ContainerDied","Data":"a8798d6563df637eedce27d76b936d9ccf5e53dcb249d23809ead3be7627744f"} Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.687893 4945 scope.go:117] "RemoveContainer" containerID="310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.689344 4945 generic.go:334] "Generic (PLEG): container finished" podID="a252dc40-e13b-400d-a2b4-8b7e99d9eca1" containerID="b36f833544bf48a858c599b0ee013f7e3e8a075cde7a98be7ffec3c49972cc48" exitCode=0 Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.689390 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a252dc40-e13b-400d-a2b4-8b7e99d9eca1","Type":"ContainerDied","Data":"b36f833544bf48a858c599b0ee013f7e3e8a075cde7a98be7ffec3c49972cc48"} Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.690993 4945 generic.go:334] "Generic (PLEG): container finished" podID="e51e86c4-2107-417e-9a3d-383d7a60da48" containerID="eb70d577a2b8a769b990acc4892ed74d29111d41d0e07f0fcb2ccea30037f24f" exitCode=0 Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.691042 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e51e86c4-2107-417e-9a3d-383d7a60da48","Type":"ContainerDied","Data":"eb70d577a2b8a769b990acc4892ed74d29111d41d0e07f0fcb2ccea30037f24f"} Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.733211 4945 scope.go:117] "RemoveContainer" containerID="964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.778599 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-swift-storage-0\") pod \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.778643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-nb\") pod \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.778689 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-svc\") pod \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.778722 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-sb\") pod \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.778767 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-config\") pod \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.778898 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l6cf\" (UniqueName: \"kubernetes.io/projected/871f25c1-f5b7-4d76-94bd-e93a39fb3724-kube-api-access-5l6cf\") pod \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.778920 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-openstack-edpm-ipam\") pod \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\" (UID: \"871f25c1-f5b7-4d76-94bd-e93a39fb3724\") " Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.814538 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/871f25c1-f5b7-4d76-94bd-e93a39fb3724-kube-api-access-5l6cf" (OuterVolumeSpecName: "kube-api-access-5l6cf") pod "871f25c1-f5b7-4d76-94bd-e93a39fb3724" (UID: "871f25c1-f5b7-4d76-94bd-e93a39fb3724"). InnerVolumeSpecName "kube-api-access-5l6cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.898632 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l6cf\" (UniqueName: \"kubernetes.io/projected/871f25c1-f5b7-4d76-94bd-e93a39fb3724-kube-api-access-5l6cf\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.908678 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "871f25c1-f5b7-4d76-94bd-e93a39fb3724" (UID: "871f25c1-f5b7-4d76-94bd-e93a39fb3724"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.911522 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "871f25c1-f5b7-4d76-94bd-e93a39fb3724" (UID: "871f25c1-f5b7-4d76-94bd-e93a39fb3724"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.919363 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-config" (OuterVolumeSpecName: "config") pod "871f25c1-f5b7-4d76-94bd-e93a39fb3724" (UID: "871f25c1-f5b7-4d76-94bd-e93a39fb3724"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.930303 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "871f25c1-f5b7-4d76-94bd-e93a39fb3724" (UID: "871f25c1-f5b7-4d76-94bd-e93a39fb3724"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.940972 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "871f25c1-f5b7-4d76-94bd-e93a39fb3724" (UID: "871f25c1-f5b7-4d76-94bd-e93a39fb3724"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:16 crc kubenswrapper[4945]: I1008 15:39:16.948746 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "871f25c1-f5b7-4d76-94bd-e93a39fb3724" (UID: "871f25c1-f5b7-4d76-94bd-e93a39fb3724"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:16.999979 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.000008 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.000018 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.000026 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.000034 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.000043 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/871f25c1-f5b7-4d76-94bd-e93a39fb3724-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.034686 4945 scope.go:117] "RemoveContainer" containerID="310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc" Oct 08 15:39:17 crc kubenswrapper[4945]: E1008 15:39:17.035044 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc\": container with ID starting with 310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc not found: ID does not exist" containerID="310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.035069 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc"} err="failed to get container status \"310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc\": rpc error: code = NotFound desc = could not find container \"310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc\": container with ID starting with 310ad5ac7055448751027a6654bf4f0f7dbb91393a293b29fc16c364299d2cfc not found: ID does not exist" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.035087 4945 scope.go:117] "RemoveContainer" containerID="964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396" Oct 08 15:39:17 crc kubenswrapper[4945]: E1008 15:39:17.035412 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396\": container with ID starting with 964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396 not found: ID does not exist" containerID="964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.035429 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396"} err="failed to get container status \"964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396\": rpc error: code = NotFound desc = could not find container \"964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396\": container with ID starting with 964069ec9efb7c4d6a537194e1f0ea69089d7c2a8f5fc62d8b14e304ba63c396 not found: ID does not exist" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.702124 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b8889b6c9-m9r4w" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.704705 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a252dc40-e13b-400d-a2b4-8b7e99d9eca1","Type":"ContainerStarted","Data":"587664f4a622a231909fe74af86efe7771d9d1c90ca0b11339d146881da6e296"} Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.705185 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.706797 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e51e86c4-2107-417e-9a3d-383d7a60da48","Type":"ContainerStarted","Data":"ca59b29fb9d63c1d053694356e5f0df59d8723d0a235bcc0e038136e16e6a50b"} Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.707200 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.729419 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.729405402 podStartE2EDuration="36.729405402s" podCreationTimestamp="2025-10-08 15:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:39:17.728045179 +0000 UTC m=+1567.081960120" watchObservedRunningTime="2025-10-08 15:39:17.729405402 +0000 UTC m=+1567.083320304" Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.756869 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b8889b6c9-m9r4w"] Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.766177 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b8889b6c9-m9r4w"] Oct 08 15:39:17 crc kubenswrapper[4945]: I1008 15:39:17.783333 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.783313309 podStartE2EDuration="36.783313309s" podCreationTimestamp="2025-10-08 15:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:39:17.77685509 +0000 UTC m=+1567.130769981" watchObservedRunningTime="2025-10-08 15:39:17.783313309 +0000 UTC m=+1567.137228210" Oct 08 15:39:18 crc kubenswrapper[4945]: I1008 15:39:18.035142 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" path="/var/lib/kubelet/pods/871f25c1-f5b7-4d76-94bd-e93a39fb3724/volumes" Oct 08 15:39:18 crc kubenswrapper[4945]: I1008 15:39:18.720796 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmlfk" event={"ID":"271df054-2fac-48c5-aa11-898d24437cb3","Type":"ContainerDied","Data":"dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94"} Oct 08 15:39:18 crc kubenswrapper[4945]: I1008 15:39:18.720733 4945 generic.go:334] "Generic (PLEG): container finished" podID="271df054-2fac-48c5-aa11-898d24437cb3" containerID="dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94" exitCode=0 Oct 08 15:39:19 crc kubenswrapper[4945]: I1008 15:39:19.184913 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:39:19 crc kubenswrapper[4945]: I1008 15:39:19.185660 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:39:20 crc kubenswrapper[4945]: I1008 15:39:20.748191 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmlfk" event={"ID":"271df054-2fac-48c5-aa11-898d24437cb3","Type":"ContainerStarted","Data":"64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460"} Oct 08 15:39:20 crc kubenswrapper[4945]: I1008 15:39:20.774864 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bmlfk" podStartSLOduration=2.780600393 podStartE2EDuration="7.77483381s" podCreationTimestamp="2025-10-08 15:39:13 +0000 UTC" firstStartedPulling="2025-10-08 15:39:14.665652104 +0000 UTC m=+1564.019567045" lastFinishedPulling="2025-10-08 15:39:19.659885561 +0000 UTC m=+1569.013800462" observedRunningTime="2025-10-08 15:39:20.76705644 +0000 UTC m=+1570.120971391" watchObservedRunningTime="2025-10-08 15:39:20.77483381 +0000 UTC m=+1570.128748731" Oct 08 15:39:23 crc kubenswrapper[4945]: I1008 15:39:23.502802 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:23 crc kubenswrapper[4945]: I1008 15:39:23.503540 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:23 crc kubenswrapper[4945]: I1008 15:39:23.577501 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:31 crc kubenswrapper[4945]: I1008 15:39:31.767280 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 15:39:31 crc kubenswrapper[4945]: I1008 15:39:31.836409 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e51e86c4-2107-417e-9a3d-383d7a60da48" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.228:5671: connect: connection refused" Oct 08 15:39:33 crc kubenswrapper[4945]: I1008 15:39:33.556638 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:33 crc kubenswrapper[4945]: I1008 15:39:33.611166 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bmlfk"] Oct 08 15:39:33 crc kubenswrapper[4945]: I1008 15:39:33.885384 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bmlfk" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="registry-server" containerID="cri-o://64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460" gracePeriod=2 Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.410927 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.442921 4945 scope.go:117] "RemoveContainer" containerID="7cbbba4d77105c1ea7fdaa1f04b108f3689c6121e56befb96a2bc2ecb6f323cf" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.466025 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9vh9\" (UniqueName: \"kubernetes.io/projected/271df054-2fac-48c5-aa11-898d24437cb3-kube-api-access-c9vh9\") pod \"271df054-2fac-48c5-aa11-898d24437cb3\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.466223 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-utilities\") pod \"271df054-2fac-48c5-aa11-898d24437cb3\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.466255 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-catalog-content\") pod \"271df054-2fac-48c5-aa11-898d24437cb3\" (UID: \"271df054-2fac-48c5-aa11-898d24437cb3\") " Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.473146 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-utilities" (OuterVolumeSpecName: "utilities") pod "271df054-2fac-48c5-aa11-898d24437cb3" (UID: "271df054-2fac-48c5-aa11-898d24437cb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.479802 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271df054-2fac-48c5-aa11-898d24437cb3-kube-api-access-c9vh9" (OuterVolumeSpecName: "kube-api-access-c9vh9") pod "271df054-2fac-48c5-aa11-898d24437cb3" (UID: "271df054-2fac-48c5-aa11-898d24437cb3"). InnerVolumeSpecName "kube-api-access-c9vh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.521417 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "271df054-2fac-48c5-aa11-898d24437cb3" (UID: "271df054-2fac-48c5-aa11-898d24437cb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.568651 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9vh9\" (UniqueName: \"kubernetes.io/projected/271df054-2fac-48c5-aa11-898d24437cb3-kube-api-access-c9vh9\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.568693 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.568705 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271df054-2fac-48c5-aa11-898d24437cb3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.612312 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8"] Oct 08 15:39:34 crc kubenswrapper[4945]: E1008 15:39:34.618320 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="registry-server" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.618355 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="registry-server" Oct 08 15:39:34 crc kubenswrapper[4945]: E1008 15:39:34.618379 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="extract-content" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.618388 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="extract-content" Oct 08 15:39:34 crc kubenswrapper[4945]: E1008 15:39:34.618398 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerName="init" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.618403 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerName="init" Oct 08 15:39:34 crc kubenswrapper[4945]: E1008 15:39:34.618427 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerName="dnsmasq-dns" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.618435 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerName="dnsmasq-dns" Oct 08 15:39:34 crc kubenswrapper[4945]: E1008 15:39:34.618455 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="extract-utilities" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.618462 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="extract-utilities" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.618715 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="871f25c1-f5b7-4d76-94bd-e93a39fb3724" containerName="dnsmasq-dns" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.618947 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="271df054-2fac-48c5-aa11-898d24437cb3" containerName="registry-server" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.619655 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.621773 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.622184 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.622849 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.623381 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.631324 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8"] Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.669607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.669667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.669806 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvtjm\" (UniqueName: \"kubernetes.io/projected/5acda85d-4b0e-4468-b19e-f2a13dc16935-kube-api-access-qvtjm\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.669937 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.772323 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.772391 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.772434 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvtjm\" (UniqueName: \"kubernetes.io/projected/5acda85d-4b0e-4468-b19e-f2a13dc16935-kube-api-access-qvtjm\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.772491 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.776588 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.776622 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.777774 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.792148 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvtjm\" (UniqueName: \"kubernetes.io/projected/5acda85d-4b0e-4468-b19e-f2a13dc16935-kube-api-access-qvtjm\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.901188 4945 generic.go:334] "Generic (PLEG): container finished" podID="271df054-2fac-48c5-aa11-898d24437cb3" containerID="64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460" exitCode=0 Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.901241 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmlfk" event={"ID":"271df054-2fac-48c5-aa11-898d24437cb3","Type":"ContainerDied","Data":"64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460"} Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.901299 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmlfk" event={"ID":"271df054-2fac-48c5-aa11-898d24437cb3","Type":"ContainerDied","Data":"8b59a35ecacd7b533700ecd0e92d59b36e81cb69199eaf489636f031633d0bb6"} Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.901325 4945 scope.go:117] "RemoveContainer" containerID="64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.901317 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmlfk" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.927715 4945 scope.go:117] "RemoveContainer" containerID="dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.938050 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bmlfk"] Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.948507 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bmlfk"] Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.954249 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:39:34 crc kubenswrapper[4945]: I1008 15:39:34.969231 4945 scope.go:117] "RemoveContainer" containerID="d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796" Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.019799 4945 scope.go:117] "RemoveContainer" containerID="64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460" Oct 08 15:39:35 crc kubenswrapper[4945]: E1008 15:39:35.020811 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460\": container with ID starting with 64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460 not found: ID does not exist" containerID="64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460" Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.020871 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460"} err="failed to get container status \"64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460\": rpc error: code = NotFound desc = could not find container \"64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460\": container with ID starting with 64aac8e9b9db9fffc3b126a7a2d7136106cad3a60dfd61a0169b02a89b0c3460 not found: ID does not exist" Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.020906 4945 scope.go:117] "RemoveContainer" containerID="dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94" Oct 08 15:39:35 crc kubenswrapper[4945]: E1008 15:39:35.021395 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94\": container with ID starting with dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94 not found: ID does not exist" containerID="dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94" Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.021449 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94"} err="failed to get container status \"dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94\": rpc error: code = NotFound desc = could not find container \"dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94\": container with ID starting with dedfa52e007003d2975fb343317367099acc3ebd90252f1069dd963775fb2a94 not found: ID does not exist" Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.021484 4945 scope.go:117] "RemoveContainer" containerID="d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796" Oct 08 15:39:35 crc kubenswrapper[4945]: E1008 15:39:35.021766 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796\": container with ID starting with d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796 not found: ID does not exist" containerID="d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796" Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.021803 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796"} err="failed to get container status \"d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796\": rpc error: code = NotFound desc = could not find container \"d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796\": container with ID starting with d1c92716d337f704960acaf4ddcc7701b2e53627bfe97d80ab4daabd35228796 not found: ID does not exist" Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.568353 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8"] Oct 08 15:39:35 crc kubenswrapper[4945]: W1008 15:39:35.581627 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5acda85d_4b0e_4468_b19e_f2a13dc16935.slice/crio-564e8e6f3f779bcf361e4839ef53bd1b02bb70bda95f1fd84d7d839cfd27aeea WatchSource:0}: Error finding container 564e8e6f3f779bcf361e4839ef53bd1b02bb70bda95f1fd84d7d839cfd27aeea: Status 404 returned error can't find the container with id 564e8e6f3f779bcf361e4839ef53bd1b02bb70bda95f1fd84d7d839cfd27aeea Oct 08 15:39:35 crc kubenswrapper[4945]: I1008 15:39:35.913458 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" event={"ID":"5acda85d-4b0e-4468-b19e-f2a13dc16935","Type":"ContainerStarted","Data":"564e8e6f3f779bcf361e4839ef53bd1b02bb70bda95f1fd84d7d839cfd27aeea"} Oct 08 15:39:36 crc kubenswrapper[4945]: I1008 15:39:36.039559 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="271df054-2fac-48c5-aa11-898d24437cb3" path="/var/lib/kubelet/pods/271df054-2fac-48c5-aa11-898d24437cb3/volumes" Oct 08 15:39:41 crc kubenswrapper[4945]: I1008 15:39:41.836510 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 15:39:49 crc kubenswrapper[4945]: I1008 15:39:49.184652 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:39:49 crc kubenswrapper[4945]: I1008 15:39:49.185374 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:39:49 crc kubenswrapper[4945]: I1008 15:39:49.185451 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:39:49 crc kubenswrapper[4945]: I1008 15:39:49.186416 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:39:49 crc kubenswrapper[4945]: I1008 15:39:49.186488 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" gracePeriod=600 Oct 08 15:39:50 crc kubenswrapper[4945]: I1008 15:39:50.100833 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" exitCode=0 Oct 08 15:39:50 crc kubenswrapper[4945]: I1008 15:39:50.101347 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0"} Oct 08 15:39:50 crc kubenswrapper[4945]: I1008 15:39:50.101457 4945 scope.go:117] "RemoveContainer" containerID="d5311a5a778713e57615c5afa0552244f2b5dba47b435a164e79654109f1ba68" Oct 08 15:39:52 crc kubenswrapper[4945]: E1008 15:39:52.241548 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:39:52 crc kubenswrapper[4945]: E1008 15:39:52.374348 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Oct 08 15:39:52 crc kubenswrapper[4945]: E1008 15:39:52.374553 4945 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 08 15:39:52 crc kubenswrapper[4945]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 08 15:39:52 crc kubenswrapper[4945]: - hosts: all Oct 08 15:39:52 crc kubenswrapper[4945]: strategy: linear Oct 08 15:39:52 crc kubenswrapper[4945]: tasks: Oct 08 15:39:52 crc kubenswrapper[4945]: - name: Enable podified-repos Oct 08 15:39:52 crc kubenswrapper[4945]: become: true Oct 08 15:39:52 crc kubenswrapper[4945]: ansible.builtin.shell: | Oct 08 15:39:52 crc kubenswrapper[4945]: set -euxo pipefail Oct 08 15:39:52 crc kubenswrapper[4945]: pushd /var/tmp Oct 08 15:39:52 crc kubenswrapper[4945]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Oct 08 15:39:52 crc kubenswrapper[4945]: pushd repo-setup-main Oct 08 15:39:52 crc kubenswrapper[4945]: python3 -m venv ./venv Oct 08 15:39:52 crc kubenswrapper[4945]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Oct 08 15:39:52 crc kubenswrapper[4945]: ./venv/bin/repo-setup current-podified -b antelope Oct 08 15:39:52 crc kubenswrapper[4945]: popd Oct 08 15:39:52 crc kubenswrapper[4945]: rm -rf repo-setup-main Oct 08 15:39:52 crc kubenswrapper[4945]: Oct 08 15:39:52 crc kubenswrapper[4945]: Oct 08 15:39:52 crc kubenswrapper[4945]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 08 15:39:52 crc kubenswrapper[4945]: edpm_override_hosts: openstack-edpm-ipam Oct 08 15:39:52 crc kubenswrapper[4945]: edpm_service_type: repo-setup Oct 08 15:39:52 crc kubenswrapper[4945]: Oct 08 15:39:52 crc kubenswrapper[4945]: Oct 08 15:39:52 crc kubenswrapper[4945]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qvtjm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8_openstack(5acda85d-4b0e-4468-b19e-f2a13dc16935): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 08 15:39:52 crc kubenswrapper[4945]: > logger="UnhandledError" Oct 08 15:39:52 crc kubenswrapper[4945]: E1008 15:39:52.376168 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" podUID="5acda85d-4b0e-4468-b19e-f2a13dc16935" Oct 08 15:39:53 crc kubenswrapper[4945]: I1008 15:39:53.138508 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:39:53 crc kubenswrapper[4945]: E1008 15:39:53.138899 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:39:53 crc kubenswrapper[4945]: E1008 15:39:53.138984 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" podUID="5acda85d-4b0e-4468-b19e-f2a13dc16935" Oct 08 15:40:04 crc kubenswrapper[4945]: I1008 15:40:04.027021 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:40:04 crc kubenswrapper[4945]: E1008 15:40:04.029692 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:40:07 crc kubenswrapper[4945]: I1008 15:40:07.029817 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:40:09 crc kubenswrapper[4945]: I1008 15:40:09.298274 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" event={"ID":"5acda85d-4b0e-4468-b19e-f2a13dc16935","Type":"ContainerStarted","Data":"534bcce6055b6553f6fc740065382458a6a0022f7b7adb9015f2f29bce1ac564"} Oct 08 15:40:09 crc kubenswrapper[4945]: I1008 15:40:09.328590 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" podStartSLOduration=1.958223008 podStartE2EDuration="35.328563045s" podCreationTimestamp="2025-10-08 15:39:34 +0000 UTC" firstStartedPulling="2025-10-08 15:39:35.585707907 +0000 UTC m=+1584.939622808" lastFinishedPulling="2025-10-08 15:40:08.956047944 +0000 UTC m=+1618.309962845" observedRunningTime="2025-10-08 15:40:09.314738265 +0000 UTC m=+1618.668653186" watchObservedRunningTime="2025-10-08 15:40:09.328563045 +0000 UTC m=+1618.682477986" Oct 08 15:40:17 crc kubenswrapper[4945]: I1008 15:40:17.024138 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:40:17 crc kubenswrapper[4945]: E1008 15:40:17.024796 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:40:24 crc kubenswrapper[4945]: I1008 15:40:24.459226 4945 generic.go:334] "Generic (PLEG): container finished" podID="5acda85d-4b0e-4468-b19e-f2a13dc16935" containerID="534bcce6055b6553f6fc740065382458a6a0022f7b7adb9015f2f29bce1ac564" exitCode=0 Oct 08 15:40:24 crc kubenswrapper[4945]: I1008 15:40:24.459380 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" event={"ID":"5acda85d-4b0e-4468-b19e-f2a13dc16935","Type":"ContainerDied","Data":"534bcce6055b6553f6fc740065382458a6a0022f7b7adb9015f2f29bce1ac564"} Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.145468 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.244099 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvtjm\" (UniqueName: \"kubernetes.io/projected/5acda85d-4b0e-4468-b19e-f2a13dc16935-kube-api-access-qvtjm\") pod \"5acda85d-4b0e-4468-b19e-f2a13dc16935\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.244269 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-ssh-key\") pod \"5acda85d-4b0e-4468-b19e-f2a13dc16935\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.244448 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-inventory\") pod \"5acda85d-4b0e-4468-b19e-f2a13dc16935\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.244569 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-repo-setup-combined-ca-bundle\") pod \"5acda85d-4b0e-4468-b19e-f2a13dc16935\" (UID: \"5acda85d-4b0e-4468-b19e-f2a13dc16935\") " Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.251040 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5acda85d-4b0e-4468-b19e-f2a13dc16935-kube-api-access-qvtjm" (OuterVolumeSpecName: "kube-api-access-qvtjm") pod "5acda85d-4b0e-4468-b19e-f2a13dc16935" (UID: "5acda85d-4b0e-4468-b19e-f2a13dc16935"). InnerVolumeSpecName "kube-api-access-qvtjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.251371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5acda85d-4b0e-4468-b19e-f2a13dc16935" (UID: "5acda85d-4b0e-4468-b19e-f2a13dc16935"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.287983 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-inventory" (OuterVolumeSpecName: "inventory") pod "5acda85d-4b0e-4468-b19e-f2a13dc16935" (UID: "5acda85d-4b0e-4468-b19e-f2a13dc16935"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.289527 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5acda85d-4b0e-4468-b19e-f2a13dc16935" (UID: "5acda85d-4b0e-4468-b19e-f2a13dc16935"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.347451 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.347491 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.347506 4945 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5acda85d-4b0e-4468-b19e-f2a13dc16935-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.347521 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvtjm\" (UniqueName: \"kubernetes.io/projected/5acda85d-4b0e-4468-b19e-f2a13dc16935-kube-api-access-qvtjm\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.488045 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" event={"ID":"5acda85d-4b0e-4468-b19e-f2a13dc16935","Type":"ContainerDied","Data":"564e8e6f3f779bcf361e4839ef53bd1b02bb70bda95f1fd84d7d839cfd27aeea"} Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.488143 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="564e8e6f3f779bcf361e4839ef53bd1b02bb70bda95f1fd84d7d839cfd27aeea" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.488272 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.583022 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl"] Oct 08 15:40:26 crc kubenswrapper[4945]: E1008 15:40:26.583470 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5acda85d-4b0e-4468-b19e-f2a13dc16935" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.583493 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5acda85d-4b0e-4468-b19e-f2a13dc16935" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.583776 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5acda85d-4b0e-4468-b19e-f2a13dc16935" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.584482 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.586597 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.586743 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.590579 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.590972 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.596142 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl"] Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.756295 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.756344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpl5p\" (UniqueName: \"kubernetes.io/projected/9afefec4-1f33-4022-9f2d-58f7e48e90f8-kube-api-access-vpl5p\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.756376 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.857791 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.858047 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpl5p\" (UniqueName: \"kubernetes.io/projected/9afefec4-1f33-4022-9f2d-58f7e48e90f8-kube-api-access-vpl5p\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.858187 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.862740 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.862813 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.876474 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpl5p\" (UniqueName: \"kubernetes.io/projected/9afefec4-1f33-4022-9f2d-58f7e48e90f8-kube-api-access-vpl5p\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2sgnl\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:26 crc kubenswrapper[4945]: I1008 15:40:26.903902 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:27 crc kubenswrapper[4945]: I1008 15:40:27.448701 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl"] Oct 08 15:40:27 crc kubenswrapper[4945]: W1008 15:40:27.450004 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9afefec4_1f33_4022_9f2d_58f7e48e90f8.slice/crio-cc3c9130c6c0f198a0edf26dd8bd87c6f031bf6ffadb8182c31dfdf8412cfcdd WatchSource:0}: Error finding container cc3c9130c6c0f198a0edf26dd8bd87c6f031bf6ffadb8182c31dfdf8412cfcdd: Status 404 returned error can't find the container with id cc3c9130c6c0f198a0edf26dd8bd87c6f031bf6ffadb8182c31dfdf8412cfcdd Oct 08 15:40:27 crc kubenswrapper[4945]: I1008 15:40:27.504467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" event={"ID":"9afefec4-1f33-4022-9f2d-58f7e48e90f8","Type":"ContainerStarted","Data":"cc3c9130c6c0f198a0edf26dd8bd87c6f031bf6ffadb8182c31dfdf8412cfcdd"} Oct 08 15:40:29 crc kubenswrapper[4945]: I1008 15:40:29.534318 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" event={"ID":"9afefec4-1f33-4022-9f2d-58f7e48e90f8","Type":"ContainerStarted","Data":"d2b671d60dd38fc0df432dbc6890186ce8a55bcdb8a798f1216fb15cf17c5317"} Oct 08 15:40:29 crc kubenswrapper[4945]: I1008 15:40:29.554072 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" podStartSLOduration=2.7244048100000002 podStartE2EDuration="3.554041594s" podCreationTimestamp="2025-10-08 15:40:26 +0000 UTC" firstStartedPulling="2025-10-08 15:40:27.453678059 +0000 UTC m=+1636.807592950" lastFinishedPulling="2025-10-08 15:40:28.283314803 +0000 UTC m=+1637.637229734" observedRunningTime="2025-10-08 15:40:29.551586204 +0000 UTC m=+1638.905501125" watchObservedRunningTime="2025-10-08 15:40:29.554041594 +0000 UTC m=+1638.907956535" Oct 08 15:40:31 crc kubenswrapper[4945]: I1008 15:40:31.025048 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:40:31 crc kubenswrapper[4945]: E1008 15:40:31.025641 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:40:31 crc kubenswrapper[4945]: I1008 15:40:31.561420 4945 generic.go:334] "Generic (PLEG): container finished" podID="9afefec4-1f33-4022-9f2d-58f7e48e90f8" containerID="d2b671d60dd38fc0df432dbc6890186ce8a55bcdb8a798f1216fb15cf17c5317" exitCode=0 Oct 08 15:40:31 crc kubenswrapper[4945]: I1008 15:40:31.561469 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" event={"ID":"9afefec4-1f33-4022-9f2d-58f7e48e90f8","Type":"ContainerDied","Data":"d2b671d60dd38fc0df432dbc6890186ce8a55bcdb8a798f1216fb15cf17c5317"} Oct 08 15:40:32 crc kubenswrapper[4945]: I1008 15:40:32.968529 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.088411 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpl5p\" (UniqueName: \"kubernetes.io/projected/9afefec4-1f33-4022-9f2d-58f7e48e90f8-kube-api-access-vpl5p\") pod \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.089303 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-inventory\") pod \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.089561 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-ssh-key\") pod \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\" (UID: \"9afefec4-1f33-4022-9f2d-58f7e48e90f8\") " Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.093450 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9afefec4-1f33-4022-9f2d-58f7e48e90f8-kube-api-access-vpl5p" (OuterVolumeSpecName: "kube-api-access-vpl5p") pod "9afefec4-1f33-4022-9f2d-58f7e48e90f8" (UID: "9afefec4-1f33-4022-9f2d-58f7e48e90f8"). InnerVolumeSpecName "kube-api-access-vpl5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.115424 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-inventory" (OuterVolumeSpecName: "inventory") pod "9afefec4-1f33-4022-9f2d-58f7e48e90f8" (UID: "9afefec4-1f33-4022-9f2d-58f7e48e90f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.135295 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9afefec4-1f33-4022-9f2d-58f7e48e90f8" (UID: "9afefec4-1f33-4022-9f2d-58f7e48e90f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.192237 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.192320 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpl5p\" (UniqueName: \"kubernetes.io/projected/9afefec4-1f33-4022-9f2d-58f7e48e90f8-kube-api-access-vpl5p\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.192339 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9afefec4-1f33-4022-9f2d-58f7e48e90f8-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.584556 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" event={"ID":"9afefec4-1f33-4022-9f2d-58f7e48e90f8","Type":"ContainerDied","Data":"cc3c9130c6c0f198a0edf26dd8bd87c6f031bf6ffadb8182c31dfdf8412cfcdd"} Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.584598 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc3c9130c6c0f198a0edf26dd8bd87c6f031bf6ffadb8182c31dfdf8412cfcdd" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.584631 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2sgnl" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.660829 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2"] Oct 08 15:40:33 crc kubenswrapper[4945]: E1008 15:40:33.661443 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9afefec4-1f33-4022-9f2d-58f7e48e90f8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.661477 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9afefec4-1f33-4022-9f2d-58f7e48e90f8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.661827 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9afefec4-1f33-4022-9f2d-58f7e48e90f8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.662926 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.666062 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.666518 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.666812 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.667099 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.675218 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2"] Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.805372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.806467 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.806640 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn2hc\" (UniqueName: \"kubernetes.io/projected/de2e46b1-591e-4540-b69e-eee056133019-kube-api-access-cn2hc\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.806749 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.909478 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.909612 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn2hc\" (UniqueName: \"kubernetes.io/projected/de2e46b1-591e-4540-b69e-eee056133019-kube-api-access-cn2hc\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.909923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.910361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.916458 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.917929 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.918340 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.934312 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn2hc\" (UniqueName: \"kubernetes.io/projected/de2e46b1-591e-4540-b69e-eee056133019-kube-api-access-cn2hc\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:33 crc kubenswrapper[4945]: I1008 15:40:33.998592 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:40:34 crc kubenswrapper[4945]: I1008 15:40:34.561253 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2"] Oct 08 15:40:34 crc kubenswrapper[4945]: W1008 15:40:34.562888 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde2e46b1_591e_4540_b69e_eee056133019.slice/crio-507e99392117846ff76dabb15a3ae31d55a9a78b5b820cce8a96986ac9aeebad WatchSource:0}: Error finding container 507e99392117846ff76dabb15a3ae31d55a9a78b5b820cce8a96986ac9aeebad: Status 404 returned error can't find the container with id 507e99392117846ff76dabb15a3ae31d55a9a78b5b820cce8a96986ac9aeebad Oct 08 15:40:34 crc kubenswrapper[4945]: I1008 15:40:34.593769 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" event={"ID":"de2e46b1-591e-4540-b69e-eee056133019","Type":"ContainerStarted","Data":"507e99392117846ff76dabb15a3ae31d55a9a78b5b820cce8a96986ac9aeebad"} Oct 08 15:40:34 crc kubenswrapper[4945]: I1008 15:40:34.611438 4945 scope.go:117] "RemoveContainer" containerID="9481b142aff7c535dd1312fc6e040c7ce6eb4d1b07e29e868578ce368cd06064" Oct 08 15:40:36 crc kubenswrapper[4945]: I1008 15:40:36.624272 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" event={"ID":"de2e46b1-591e-4540-b69e-eee056133019","Type":"ContainerStarted","Data":"36ae2bdecc00ceae8a4aac355830100ad0bd3605d7dcae22315608a38b46829f"} Oct 08 15:40:36 crc kubenswrapper[4945]: I1008 15:40:36.668445 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" podStartSLOduration=2.718898926 podStartE2EDuration="3.668411418s" podCreationTimestamp="2025-10-08 15:40:33 +0000 UTC" firstStartedPulling="2025-10-08 15:40:34.565893991 +0000 UTC m=+1643.919808892" lastFinishedPulling="2025-10-08 15:40:35.515406463 +0000 UTC m=+1644.869321384" observedRunningTime="2025-10-08 15:40:36.653278227 +0000 UTC m=+1646.007193178" watchObservedRunningTime="2025-10-08 15:40:36.668411418 +0000 UTC m=+1646.022326359" Oct 08 15:40:45 crc kubenswrapper[4945]: I1008 15:40:45.024649 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:40:45 crc kubenswrapper[4945]: E1008 15:40:45.025607 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:40:57 crc kubenswrapper[4945]: I1008 15:40:57.024160 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:40:57 crc kubenswrapper[4945]: E1008 15:40:57.024925 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:41:11 crc kubenswrapper[4945]: I1008 15:41:11.023797 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:41:11 crc kubenswrapper[4945]: E1008 15:41:11.025612 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:41:22 crc kubenswrapper[4945]: I1008 15:41:22.031406 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:41:22 crc kubenswrapper[4945]: E1008 15:41:22.032168 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:41:34 crc kubenswrapper[4945]: I1008 15:41:34.711373 4945 scope.go:117] "RemoveContainer" containerID="9e390c3243fdbc50edb03d4201efd8e7bc705d890f0f3c40703ff960ab333154" Oct 08 15:41:34 crc kubenswrapper[4945]: I1008 15:41:34.751987 4945 scope.go:117] "RemoveContainer" containerID="305d0cda5ee94e95efdbedd6f7f48b8d533c3afbcad085b02e420045d93135b3" Oct 08 15:41:36 crc kubenswrapper[4945]: I1008 15:41:36.025325 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:41:36 crc kubenswrapper[4945]: E1008 15:41:36.025942 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:41:49 crc kubenswrapper[4945]: I1008 15:41:49.024257 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:41:49 crc kubenswrapper[4945]: E1008 15:41:49.025064 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:42:04 crc kubenswrapper[4945]: I1008 15:42:04.025434 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:42:04 crc kubenswrapper[4945]: E1008 15:42:04.026503 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:42:18 crc kubenswrapper[4945]: I1008 15:42:18.024790 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:42:18 crc kubenswrapper[4945]: E1008 15:42:18.025804 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:42:31 crc kubenswrapper[4945]: I1008 15:42:31.025295 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:42:31 crc kubenswrapper[4945]: E1008 15:42:31.026302 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:42:34 crc kubenswrapper[4945]: I1008 15:42:34.843079 4945 scope.go:117] "RemoveContainer" containerID="61a9e3556abf6eba2040ec62743a6be0f8b9864d047c2a6d5017a6e1aede5f28" Oct 08 15:42:36 crc kubenswrapper[4945]: I1008 15:42:36.057613 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-n5pnx"] Oct 08 15:42:36 crc kubenswrapper[4945]: I1008 15:42:36.074856 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-n5pnx"] Oct 08 15:42:38 crc kubenswrapper[4945]: I1008 15:42:38.044353 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc1f5af-e985-4b88-bb14-2d4cd9555188" path="/var/lib/kubelet/pods/4cc1f5af-e985-4b88-bb14-2d4cd9555188/volumes" Oct 08 15:42:41 crc kubenswrapper[4945]: I1008 15:42:41.044307 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-8lkkh"] Oct 08 15:42:41 crc kubenswrapper[4945]: I1008 15:42:41.054984 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-8lkkh"] Oct 08 15:42:42 crc kubenswrapper[4945]: I1008 15:42:42.039501 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:42:42 crc kubenswrapper[4945]: E1008 15:42:42.040227 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:42:42 crc kubenswrapper[4945]: I1008 15:42:42.041448 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="817d5ad4-17fa-4845-9a91-c96e271d4d40" path="/var/lib/kubelet/pods/817d5ad4-17fa-4845-9a91-c96e271d4d40/volumes" Oct 08 15:42:45 crc kubenswrapper[4945]: I1008 15:42:45.036164 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-8z2zf"] Oct 08 15:42:45 crc kubenswrapper[4945]: I1008 15:42:45.044178 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-8z2zf"] Oct 08 15:42:46 crc kubenswrapper[4945]: I1008 15:42:46.039131 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9e17266-40ce-4c98-8f15-80e8da63774d" path="/var/lib/kubelet/pods/b9e17266-40ce-4c98-8f15-80e8da63774d/volumes" Oct 08 15:42:56 crc kubenswrapper[4945]: I1008 15:42:56.024971 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:42:56 crc kubenswrapper[4945]: E1008 15:42:56.025896 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:42:56 crc kubenswrapper[4945]: I1008 15:42:56.038623 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-e917-account-create-5nmhp"] Oct 08 15:42:56 crc kubenswrapper[4945]: I1008 15:42:56.044432 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f58f-account-create-t2jxv"] Oct 08 15:42:56 crc kubenswrapper[4945]: I1008 15:42:56.056228 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f58f-account-create-t2jxv"] Oct 08 15:42:56 crc kubenswrapper[4945]: I1008 15:42:56.069124 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-e917-account-create-5nmhp"] Oct 08 15:42:57 crc kubenswrapper[4945]: I1008 15:42:57.034393 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-8ae8-account-create-655qj"] Oct 08 15:42:57 crc kubenswrapper[4945]: I1008 15:42:57.042084 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-8ae8-account-create-655qj"] Oct 08 15:42:58 crc kubenswrapper[4945]: I1008 15:42:58.044296 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62fa8723-fb82-45bf-9750-2a167cdec803" path="/var/lib/kubelet/pods/62fa8723-fb82-45bf-9750-2a167cdec803/volumes" Oct 08 15:42:58 crc kubenswrapper[4945]: I1008 15:42:58.046270 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2403ece-f576-45c1-b7cd-0547eb470c75" path="/var/lib/kubelet/pods/b2403ece-f576-45c1-b7cd-0547eb470c75/volumes" Oct 08 15:42:58 crc kubenswrapper[4945]: I1008 15:42:58.047367 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b99dc5f0-37b6-45ed-8a7c-8b2299405792" path="/var/lib/kubelet/pods/b99dc5f0-37b6-45ed-8a7c-8b2299405792/volumes" Oct 08 15:43:07 crc kubenswrapper[4945]: I1008 15:43:07.024441 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:43:07 crc kubenswrapper[4945]: E1008 15:43:07.025224 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.058465 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-gzvqb"] Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.069920 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-gp4z9"] Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.081705 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-gzvqb"] Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.092595 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-wlvmx"] Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.102181 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-42dgd"] Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.111402 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-gp4z9"] Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.119515 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-wlvmx"] Oct 08 15:43:15 crc kubenswrapper[4945]: I1008 15:43:15.127639 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-42dgd"] Oct 08 15:43:16 crc kubenswrapper[4945]: I1008 15:43:16.038024 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e2edcd-fb7f-4c8b-aac6-7c6c2769e579" path="/var/lib/kubelet/pods/09e2edcd-fb7f-4c8b-aac6-7c6c2769e579/volumes" Oct 08 15:43:16 crc kubenswrapper[4945]: I1008 15:43:16.039248 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c95dc54-6062-4f5f-8ffa-79cd1a23749b" path="/var/lib/kubelet/pods/2c95dc54-6062-4f5f-8ffa-79cd1a23749b/volumes" Oct 08 15:43:16 crc kubenswrapper[4945]: I1008 15:43:16.039797 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="549f6009-8c9d-4fbb-890d-236b5c93c3fa" path="/var/lib/kubelet/pods/549f6009-8c9d-4fbb-890d-236b5c93c3fa/volumes" Oct 08 15:43:16 crc kubenswrapper[4945]: I1008 15:43:16.040984 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79250da1-c44a-4d62-a5e8-1ce671a336a2" path="/var/lib/kubelet/pods/79250da1-c44a-4d62-a5e8-1ce671a336a2/volumes" Oct 08 15:43:18 crc kubenswrapper[4945]: I1008 15:43:18.024629 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:43:18 crc kubenswrapper[4945]: E1008 15:43:18.025477 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:43:27 crc kubenswrapper[4945]: I1008 15:43:27.542866 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7cfddd55fc-6xk6v" podUID="2e5af0dd-db9f-40a7-941c-1c08c0bc5be4" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 08 15:43:32 crc kubenswrapper[4945]: I1008 15:43:32.031278 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:43:32 crc kubenswrapper[4945]: E1008 15:43:32.032209 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:43:34 crc kubenswrapper[4945]: I1008 15:43:34.906085 4945 scope.go:117] "RemoveContainer" containerID="bddda164933162b2371f0930e6e673b2c885adf33d9cb1e1876d9e9843580c57" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.052916 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-80a1-account-create-gzqfc"] Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.063598 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-80a1-account-create-gzqfc"] Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.257744 4945 scope.go:117] "RemoveContainer" containerID="3f0516edf2b32c59f0674abb53b6e769f259fcaf0649fd32cdaca1e406c2075e" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.294526 4945 scope.go:117] "RemoveContainer" containerID="3c092b22a02b2318946c47e8e94c357b173c1b21c2aaa6acd1ca8c1b54c31c47" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.432866 4945 scope.go:117] "RemoveContainer" containerID="3bb8cd25df2313b3f4ae8d27722fd83a2e72e6de26a37b51a3900ca4dda4111b" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.676317 4945 scope.go:117] "RemoveContainer" containerID="ee3647dd523a3881a0a0add10110fa5a1b57bca8f3ea7e735eff6f6a98d93b03" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.712287 4945 scope.go:117] "RemoveContainer" containerID="48df158afb6810f473c1638e7e7ae4e866fc706660c783a9be5d2ad845c7003f" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.747102 4945 scope.go:117] "RemoveContainer" containerID="011272c4b0c1eac555dfdae3036d6f495a71bbf710690bcd474820004ddb89b9" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.814146 4945 scope.go:117] "RemoveContainer" containerID="aa1eb389650b5be13d90a8c5e975c9718a381bd539f4a8ed49626d1e404fef5e" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.845343 4945 scope.go:117] "RemoveContainer" containerID="0ef9ba4567aa1ef7ddf1c20bafc2284544ee1b1bd89997c1b290f41a859c336e" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.873163 4945 scope.go:117] "RemoveContainer" containerID="bd261f78913c8998d8fcce217c2a1bdd1aae8f390bf628ebeb308940199ab21b" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.893397 4945 scope.go:117] "RemoveContainer" containerID="c7fc8b011611ab6a3d06f1830a19fdb5489c5392921cac4a36df353c4271e3ba" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.924222 4945 scope.go:117] "RemoveContainer" containerID="db3fad6c030c3ba54c04dd337f0c8d6e47e0329111210fd589b34edee2e5ab0d" Oct 08 15:43:35 crc kubenswrapper[4945]: I1008 15:43:35.985009 4945 scope.go:117] "RemoveContainer" containerID="4f845e6f03da20b18248928b9a4c30725d8c04becebef9b60f7c42e2475823ac" Oct 08 15:43:36 crc kubenswrapper[4945]: I1008 15:43:36.046892 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a113f9df-ecb0-48a1-9b10-05df2fea24c7" path="/var/lib/kubelet/pods/a113f9df-ecb0-48a1-9b10-05df2fea24c7/volumes" Oct 08 15:43:36 crc kubenswrapper[4945]: I1008 15:43:36.047563 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e695-account-create-tnzkz"] Oct 08 15:43:36 crc kubenswrapper[4945]: I1008 15:43:36.047593 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-37c5-account-create-dwc7q"] Oct 08 15:43:36 crc kubenswrapper[4945]: I1008 15:43:36.056026 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-1ef5-account-create-zz9f8"] Oct 08 15:43:36 crc kubenswrapper[4945]: I1008 15:43:36.063777 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-37c5-account-create-dwc7q"] Oct 08 15:43:36 crc kubenswrapper[4945]: I1008 15:43:36.070346 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-1ef5-account-create-zz9f8"] Oct 08 15:43:36 crc kubenswrapper[4945]: I1008 15:43:36.077029 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e695-account-create-tnzkz"] Oct 08 15:43:38 crc kubenswrapper[4945]: I1008 15:43:38.041239 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0462b93b-24b8-42a6-81cf-30e0b0b09eb7" path="/var/lib/kubelet/pods/0462b93b-24b8-42a6-81cf-30e0b0b09eb7/volumes" Oct 08 15:43:38 crc kubenswrapper[4945]: I1008 15:43:38.043189 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="269baca3-654a-415a-a631-258a2bc03e03" path="/var/lib/kubelet/pods/269baca3-654a-415a-a631-258a2bc03e03/volumes" Oct 08 15:43:38 crc kubenswrapper[4945]: I1008 15:43:38.043894 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9316cb7-a351-4ca8-8ea9-e91fd507ea18" path="/var/lib/kubelet/pods/e9316cb7-a351-4ca8-8ea9-e91fd507ea18/volumes" Oct 08 15:43:44 crc kubenswrapper[4945]: I1008 15:43:44.024609 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:43:44 crc kubenswrapper[4945]: E1008 15:43:44.025439 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:43:56 crc kubenswrapper[4945]: I1008 15:43:56.050933 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-978m6"] Oct 08 15:43:56 crc kubenswrapper[4945]: I1008 15:43:56.061835 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-978m6"] Oct 08 15:43:58 crc kubenswrapper[4945]: I1008 15:43:58.025462 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:43:58 crc kubenswrapper[4945]: E1008 15:43:58.026067 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:43:58 crc kubenswrapper[4945]: I1008 15:43:58.038742 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d7796c3-c17f-4795-b874-d5196d0deab2" path="/var/lib/kubelet/pods/6d7796c3-c17f-4795-b874-d5196d0deab2/volumes" Oct 08 15:44:01 crc kubenswrapper[4945]: I1008 15:44:01.035228 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-t8trl"] Oct 08 15:44:01 crc kubenswrapper[4945]: I1008 15:44:01.044936 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-t8trl"] Oct 08 15:44:02 crc kubenswrapper[4945]: I1008 15:44:02.037303 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf47e507-de36-49a3-9428-7104713b4d54" path="/var/lib/kubelet/pods/cf47e507-de36-49a3-9428-7104713b4d54/volumes" Oct 08 15:44:11 crc kubenswrapper[4945]: I1008 15:44:11.024753 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:44:11 crc kubenswrapper[4945]: E1008 15:44:11.025617 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:44:13 crc kubenswrapper[4945]: I1008 15:44:13.008498 4945 generic.go:334] "Generic (PLEG): container finished" podID="de2e46b1-591e-4540-b69e-eee056133019" containerID="36ae2bdecc00ceae8a4aac355830100ad0bd3605d7dcae22315608a38b46829f" exitCode=0 Oct 08 15:44:13 crc kubenswrapper[4945]: I1008 15:44:13.008604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" event={"ID":"de2e46b1-591e-4540-b69e-eee056133019","Type":"ContainerDied","Data":"36ae2bdecc00ceae8a4aac355830100ad0bd3605d7dcae22315608a38b46829f"} Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.476814 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.611939 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-ssh-key\") pod \"de2e46b1-591e-4540-b69e-eee056133019\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.612022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn2hc\" (UniqueName: \"kubernetes.io/projected/de2e46b1-591e-4540-b69e-eee056133019-kube-api-access-cn2hc\") pod \"de2e46b1-591e-4540-b69e-eee056133019\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.612379 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-inventory\") pod \"de2e46b1-591e-4540-b69e-eee056133019\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.612436 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-bootstrap-combined-ca-bundle\") pod \"de2e46b1-591e-4540-b69e-eee056133019\" (UID: \"de2e46b1-591e-4540-b69e-eee056133019\") " Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.618019 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2e46b1-591e-4540-b69e-eee056133019-kube-api-access-cn2hc" (OuterVolumeSpecName: "kube-api-access-cn2hc") pod "de2e46b1-591e-4540-b69e-eee056133019" (UID: "de2e46b1-591e-4540-b69e-eee056133019"). InnerVolumeSpecName "kube-api-access-cn2hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.626726 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "de2e46b1-591e-4540-b69e-eee056133019" (UID: "de2e46b1-591e-4540-b69e-eee056133019"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.649650 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-inventory" (OuterVolumeSpecName: "inventory") pod "de2e46b1-591e-4540-b69e-eee056133019" (UID: "de2e46b1-591e-4540-b69e-eee056133019"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.664999 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de2e46b1-591e-4540-b69e-eee056133019" (UID: "de2e46b1-591e-4540-b69e-eee056133019"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.715211 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.715258 4945 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.715274 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de2e46b1-591e-4540-b69e-eee056133019-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:14 crc kubenswrapper[4945]: I1008 15:44:14.715288 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn2hc\" (UniqueName: \"kubernetes.io/projected/de2e46b1-591e-4540-b69e-eee056133019-kube-api-access-cn2hc\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.038335 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" event={"ID":"de2e46b1-591e-4540-b69e-eee056133019","Type":"ContainerDied","Data":"507e99392117846ff76dabb15a3ae31d55a9a78b5b820cce8a96986ac9aeebad"} Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.038777 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="507e99392117846ff76dabb15a3ae31d55a9a78b5b820cce8a96986ac9aeebad" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.038403 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.131198 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8"] Oct 08 15:44:15 crc kubenswrapper[4945]: E1008 15:44:15.131597 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2e46b1-591e-4540-b69e-eee056133019" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.131615 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2e46b1-591e-4540-b69e-eee056133019" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.131807 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2e46b1-591e-4540-b69e-eee056133019" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.132472 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.134478 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.134571 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.134663 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.135516 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.146911 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8"] Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.227160 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.227344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4t5k\" (UniqueName: \"kubernetes.io/projected/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-kube-api-access-h4t5k\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.227410 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.329819 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.330029 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.330196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4t5k\" (UniqueName: \"kubernetes.io/projected/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-kube-api-access-h4t5k\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.335571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.337850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.348696 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4t5k\" (UniqueName: \"kubernetes.io/projected/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-kube-api-access-h4t5k\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:15 crc kubenswrapper[4945]: I1008 15:44:15.458690 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:44:16 crc kubenswrapper[4945]: I1008 15:44:16.010182 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8"] Oct 08 15:44:16 crc kubenswrapper[4945]: W1008 15:44:16.020098 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode59943cd_9572_4e6b_8a58_c0ec1283b2aa.slice/crio-81d74b6232e0f9bd61789440c6ee8498f4aace429d9eb70ee5ff754f5b02d1a4 WatchSource:0}: Error finding container 81d74b6232e0f9bd61789440c6ee8498f4aace429d9eb70ee5ff754f5b02d1a4: Status 404 returned error can't find the container with id 81d74b6232e0f9bd61789440c6ee8498f4aace429d9eb70ee5ff754f5b02d1a4 Oct 08 15:44:16 crc kubenswrapper[4945]: I1008 15:44:16.050704 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" event={"ID":"e59943cd-9572-4e6b-8a58-c0ec1283b2aa","Type":"ContainerStarted","Data":"81d74b6232e0f9bd61789440c6ee8498f4aace429d9eb70ee5ff754f5b02d1a4"} Oct 08 15:44:18 crc kubenswrapper[4945]: I1008 15:44:18.068412 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" event={"ID":"e59943cd-9572-4e6b-8a58-c0ec1283b2aa","Type":"ContainerStarted","Data":"14982dc341d12f6481e386bdbafc6f25c6978e93ccfe1e1dd5d68fcddbad447e"} Oct 08 15:44:19 crc kubenswrapper[4945]: I1008 15:44:19.094769 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" podStartSLOduration=2.6053880940000003 podStartE2EDuration="4.094746727s" podCreationTimestamp="2025-10-08 15:44:15 +0000 UTC" firstStartedPulling="2025-10-08 15:44:16.02263024 +0000 UTC m=+1865.376545141" lastFinishedPulling="2025-10-08 15:44:17.511988873 +0000 UTC m=+1866.865903774" observedRunningTime="2025-10-08 15:44:19.093418844 +0000 UTC m=+1868.447333775" watchObservedRunningTime="2025-10-08 15:44:19.094746727 +0000 UTC m=+1868.448661638" Oct 08 15:44:23 crc kubenswrapper[4945]: I1008 15:44:23.024524 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:44:23 crc kubenswrapper[4945]: E1008 15:44:23.025624 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:44:34 crc kubenswrapper[4945]: I1008 15:44:34.025134 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:44:34 crc kubenswrapper[4945]: E1008 15:44:34.025886 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.252784 4945 scope.go:117] "RemoveContainer" containerID="8b220812c914192ec6cd78a4b415bb6f300201396cff210462ed3c4336553bed" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.310416 4945 scope.go:117] "RemoveContainer" containerID="d05d6192f8559971e4cf7034bd5fd3505c6dc3343a8d581776f7a9847d8aa77a" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.383891 4945 scope.go:117] "RemoveContainer" containerID="b027c7df84247536b2b76ba5946c3242f6ae1576ffde04680aeaeff006508020" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.440079 4945 scope.go:117] "RemoveContainer" containerID="0ccf493da6ed3153d0181e26501441b476f23b3c48a6eb8d520752e3766b822f" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.502645 4945 scope.go:117] "RemoveContainer" containerID="7944122f426088debc84d17e6bcf7fcb196e3fec9651fab7be86ccee136061e8" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.537324 4945 scope.go:117] "RemoveContainer" containerID="7cd84effde7b665808233b56e0a46d433135176c58a1ca21c3db2500847bb495" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.558242 4945 scope.go:117] "RemoveContainer" containerID="6ec53bed8a5ddd4fbd36c6e1b8be7a4f7e4b9a6612f1176c25c5f524b6f1110a" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.607539 4945 scope.go:117] "RemoveContainer" containerID="8289e3f9eeddb02f2374bb282ac70d3e256cb55e78a8a8af423d4533a687948c" Oct 08 15:44:36 crc kubenswrapper[4945]: I1008 15:44:36.627058 4945 scope.go:117] "RemoveContainer" containerID="698be937fa5f0c9c26664708304be7fc79e96fc724ce9fee1cdbb51c8ab42c80" Oct 08 15:44:41 crc kubenswrapper[4945]: I1008 15:44:41.050972 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-qkgj2"] Oct 08 15:44:41 crc kubenswrapper[4945]: I1008 15:44:41.059872 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-qkgj2"] Oct 08 15:44:42 crc kubenswrapper[4945]: I1008 15:44:42.042606 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30020f4e-d113-4b93-baa0-553942be7a4d" path="/var/lib/kubelet/pods/30020f4e-d113-4b93-baa0-553942be7a4d/volumes" Oct 08 15:44:45 crc kubenswrapper[4945]: I1008 15:44:45.025088 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:44:45 crc kubenswrapper[4945]: E1008 15:44:45.025884 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.024974 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.166245 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s"] Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.168453 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.177584 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s"] Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.180423 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.180893 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.238126 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ea7eab-c495-4da0-bdaa-18edd47f0744-config-volume\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.238392 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdncd\" (UniqueName: \"kubernetes.io/projected/b7ea7eab-c495-4da0-bdaa-18edd47f0744-kube-api-access-mdncd\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.238566 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ea7eab-c495-4da0-bdaa-18edd47f0744-secret-volume\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.340569 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ea7eab-c495-4da0-bdaa-18edd47f0744-config-volume\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.340620 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdncd\" (UniqueName: \"kubernetes.io/projected/b7ea7eab-c495-4da0-bdaa-18edd47f0744-kube-api-access-mdncd\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.340831 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ea7eab-c495-4da0-bdaa-18edd47f0744-secret-volume\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.341929 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ea7eab-c495-4da0-bdaa-18edd47f0744-config-volume\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.349029 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ea7eab-c495-4da0-bdaa-18edd47f0744-secret-volume\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.359234 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdncd\" (UniqueName: \"kubernetes.io/projected/b7ea7eab-c495-4da0-bdaa-18edd47f0744-kube-api-access-mdncd\") pod \"collect-profiles-29332305-5fj4s\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.493508 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"00b2ab00906aa9fd3ef271df8c0b042025d08d3100676934af532ebbb72bb625"} Oct 08 15:45:00 crc kubenswrapper[4945]: I1008 15:45:00.545237 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:01 crc kubenswrapper[4945]: I1008 15:45:01.007306 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s"] Oct 08 15:45:01 crc kubenswrapper[4945]: W1008 15:45:01.010305 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7ea7eab_c495_4da0_bdaa_18edd47f0744.slice/crio-cbf8d3bd9a7a624396c338a2f465b595128cfe10c4592af189be3586e82bd1dd WatchSource:0}: Error finding container cbf8d3bd9a7a624396c338a2f465b595128cfe10c4592af189be3586e82bd1dd: Status 404 returned error can't find the container with id cbf8d3bd9a7a624396c338a2f465b595128cfe10c4592af189be3586e82bd1dd Oct 08 15:45:01 crc kubenswrapper[4945]: I1008 15:45:01.505494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" event={"ID":"b7ea7eab-c495-4da0-bdaa-18edd47f0744","Type":"ContainerStarted","Data":"5620c65ea05b581ef32a4dc80bd7defcda5a705f034c56daf6f3abdca4024dbf"} Oct 08 15:45:01 crc kubenswrapper[4945]: I1008 15:45:01.505803 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" event={"ID":"b7ea7eab-c495-4da0-bdaa-18edd47f0744","Type":"ContainerStarted","Data":"cbf8d3bd9a7a624396c338a2f465b595128cfe10c4592af189be3586e82bd1dd"} Oct 08 15:45:01 crc kubenswrapper[4945]: I1008 15:45:01.525148 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" podStartSLOduration=1.525126605 podStartE2EDuration="1.525126605s" podCreationTimestamp="2025-10-08 15:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:45:01.520398936 +0000 UTC m=+1910.874313857" watchObservedRunningTime="2025-10-08 15:45:01.525126605 +0000 UTC m=+1910.879041506" Oct 08 15:45:02 crc kubenswrapper[4945]: I1008 15:45:02.517053 4945 generic.go:334] "Generic (PLEG): container finished" podID="b7ea7eab-c495-4da0-bdaa-18edd47f0744" containerID="5620c65ea05b581ef32a4dc80bd7defcda5a705f034c56daf6f3abdca4024dbf" exitCode=0 Oct 08 15:45:02 crc kubenswrapper[4945]: I1008 15:45:02.517659 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" event={"ID":"b7ea7eab-c495-4da0-bdaa-18edd47f0744","Type":"ContainerDied","Data":"5620c65ea05b581ef32a4dc80bd7defcda5a705f034c56daf6f3abdca4024dbf"} Oct 08 15:45:03 crc kubenswrapper[4945]: I1008 15:45:03.907648 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.016783 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdncd\" (UniqueName: \"kubernetes.io/projected/b7ea7eab-c495-4da0-bdaa-18edd47f0744-kube-api-access-mdncd\") pod \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.016891 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ea7eab-c495-4da0-bdaa-18edd47f0744-secret-volume\") pod \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.016933 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ea7eab-c495-4da0-bdaa-18edd47f0744-config-volume\") pod \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\" (UID: \"b7ea7eab-c495-4da0-bdaa-18edd47f0744\") " Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.018044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7ea7eab-c495-4da0-bdaa-18edd47f0744-config-volume" (OuterVolumeSpecName: "config-volume") pod "b7ea7eab-c495-4da0-bdaa-18edd47f0744" (UID: "b7ea7eab-c495-4da0-bdaa-18edd47f0744"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.024001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7ea7eab-c495-4da0-bdaa-18edd47f0744-kube-api-access-mdncd" (OuterVolumeSpecName: "kube-api-access-mdncd") pod "b7ea7eab-c495-4da0-bdaa-18edd47f0744" (UID: "b7ea7eab-c495-4da0-bdaa-18edd47f0744"). InnerVolumeSpecName "kube-api-access-mdncd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.024134 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7ea7eab-c495-4da0-bdaa-18edd47f0744-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b7ea7eab-c495-4da0-bdaa-18edd47f0744" (UID: "b7ea7eab-c495-4da0-bdaa-18edd47f0744"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.119049 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ea7eab-c495-4da0-bdaa-18edd47f0744-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.119086 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdncd\" (UniqueName: \"kubernetes.io/projected/b7ea7eab-c495-4da0-bdaa-18edd47f0744-kube-api-access-mdncd\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.119101 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ea7eab-c495-4da0-bdaa-18edd47f0744-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.537967 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" event={"ID":"b7ea7eab-c495-4da0-bdaa-18edd47f0744","Type":"ContainerDied","Data":"cbf8d3bd9a7a624396c338a2f465b595128cfe10c4592af189be3586e82bd1dd"} Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.538330 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbf8d3bd9a7a624396c338a2f465b595128cfe10c4592af189be3586e82bd1dd" Oct 08 15:45:04 crc kubenswrapper[4945]: I1008 15:45:04.538392 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s" Oct 08 15:45:05 crc kubenswrapper[4945]: I1008 15:45:05.037163 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-s7fh4"] Oct 08 15:45:05 crc kubenswrapper[4945]: I1008 15:45:05.049931 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-s7fh4"] Oct 08 15:45:06 crc kubenswrapper[4945]: I1008 15:45:06.034306 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47794116-db67-4451-858c-bb6fea64a21c" path="/var/lib/kubelet/pods/47794116-db67-4451-858c-bb6fea64a21c/volumes" Oct 08 15:45:06 crc kubenswrapper[4945]: I1008 15:45:06.037222 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ttghn"] Oct 08 15:45:06 crc kubenswrapper[4945]: I1008 15:45:06.052379 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ttghn"] Oct 08 15:45:08 crc kubenswrapper[4945]: I1008 15:45:08.036427 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e670ed1-808d-4d6c-aa73-106476663b46" path="/var/lib/kubelet/pods/0e670ed1-808d-4d6c-aa73-106476663b46/volumes" Oct 08 15:45:14 crc kubenswrapper[4945]: I1008 15:45:14.053014 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-pfbzj"] Oct 08 15:45:14 crc kubenswrapper[4945]: I1008 15:45:14.063212 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-pfbzj"] Oct 08 15:45:16 crc kubenswrapper[4945]: I1008 15:45:16.052067 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1be99142-5236-4365-8a45-70847b54b883" path="/var/lib/kubelet/pods/1be99142-5236-4365-8a45-70847b54b883/volumes" Oct 08 15:45:23 crc kubenswrapper[4945]: I1008 15:45:23.038337 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-ztjf5"] Oct 08 15:45:23 crc kubenswrapper[4945]: I1008 15:45:23.048492 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-ztjf5"] Oct 08 15:45:24 crc kubenswrapper[4945]: I1008 15:45:24.053339 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dec4ad0-549d-49a0-9749-86b21bbaf987" path="/var/lib/kubelet/pods/3dec4ad0-549d-49a0-9749-86b21bbaf987/volumes" Oct 08 15:45:27 crc kubenswrapper[4945]: I1008 15:45:27.028922 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4bnbb"] Oct 08 15:45:27 crc kubenswrapper[4945]: I1008 15:45:27.039004 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4bnbb"] Oct 08 15:45:28 crc kubenswrapper[4945]: I1008 15:45:28.036844 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0646f7b0-6abf-436c-976f-59ab119525ad" path="/var/lib/kubelet/pods/0646f7b0-6abf-436c-976f-59ab119525ad/volumes" Oct 08 15:45:36 crc kubenswrapper[4945]: I1008 15:45:36.866841 4945 scope.go:117] "RemoveContainer" containerID="4bb443fca884f14d549c2bf017984adacea0335b8a1163dc787f72f7f492f51a" Oct 08 15:45:36 crc kubenswrapper[4945]: I1008 15:45:36.911618 4945 scope.go:117] "RemoveContainer" containerID="bea1ed470f3c1f66f159279f8b32dd1d7a731edce53ff621fa977d3372b1cac1" Oct 08 15:45:36 crc kubenswrapper[4945]: I1008 15:45:36.980220 4945 scope.go:117] "RemoveContainer" containerID="481ea462b2ffe45c8b6a5bf15ad10dc23616f5609ecafac7d281205ea1afd821" Oct 08 15:45:37 crc kubenswrapper[4945]: I1008 15:45:37.017570 4945 scope.go:117] "RemoveContainer" containerID="4f80b380f1351f165e1e0ed9bae4191556d2b142ef136edc04bb5bcfb88ba0ed" Oct 08 15:45:37 crc kubenswrapper[4945]: I1008 15:45:37.065669 4945 scope.go:117] "RemoveContainer" containerID="e5feed9bc5222c25f23c035f6293901350cf1dda4f5023e3a461ca8f7264faba" Oct 08 15:45:37 crc kubenswrapper[4945]: I1008 15:45:37.120892 4945 scope.go:117] "RemoveContainer" containerID="d51a29b99688467a7a64d40bba0f998eb8efc47d4f0e03d26b39c3dc69ab162f" Oct 08 15:45:47 crc kubenswrapper[4945]: I1008 15:45:47.058590 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-z7bqr"] Oct 08 15:45:47 crc kubenswrapper[4945]: I1008 15:45:47.068798 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5z5q5"] Oct 08 15:45:47 crc kubenswrapper[4945]: I1008 15:45:47.079097 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5z5q5"] Oct 08 15:45:47 crc kubenswrapper[4945]: I1008 15:45:47.087726 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-z7bqr"] Oct 08 15:45:48 crc kubenswrapper[4945]: I1008 15:45:48.036919 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67017489-fd71-4e43-99c6-7918bc9f41f5" path="/var/lib/kubelet/pods/67017489-fd71-4e43-99c6-7918bc9f41f5/volumes" Oct 08 15:45:48 crc kubenswrapper[4945]: I1008 15:45:48.037466 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a4bdf24-ed65-4aa7-acd9-922bd48a1b88" path="/var/lib/kubelet/pods/6a4bdf24-ed65-4aa7-acd9-922bd48a1b88/volumes" Oct 08 15:45:48 crc kubenswrapper[4945]: I1008 15:45:48.069636 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-df5bp"] Oct 08 15:45:48 crc kubenswrapper[4945]: I1008 15:45:48.081305 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-df5bp"] Oct 08 15:45:50 crc kubenswrapper[4945]: I1008 15:45:50.035062 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f69da577-0161-4756-94be-12ee983945bd" path="/var/lib/kubelet/pods/f69da577-0161-4756-94be-12ee983945bd/volumes" Oct 08 15:45:58 crc kubenswrapper[4945]: I1008 15:45:58.042041 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-95e9-account-create-z2krj"] Oct 08 15:45:58 crc kubenswrapper[4945]: I1008 15:45:58.053523 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3781-account-create-qxkhh"] Oct 08 15:45:58 crc kubenswrapper[4945]: I1008 15:45:58.064850 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f303-account-create-qwm6s"] Oct 08 15:45:58 crc kubenswrapper[4945]: I1008 15:45:58.074401 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-95e9-account-create-z2krj"] Oct 08 15:45:58 crc kubenswrapper[4945]: I1008 15:45:58.085892 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3781-account-create-qxkhh"] Oct 08 15:45:58 crc kubenswrapper[4945]: I1008 15:45:58.096435 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f303-account-create-qwm6s"] Oct 08 15:46:00 crc kubenswrapper[4945]: I1008 15:46:00.042330 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c284362-1210-4291-b285-ec21e6c862e0" path="/var/lib/kubelet/pods/5c284362-1210-4291-b285-ec21e6c862e0/volumes" Oct 08 15:46:00 crc kubenswrapper[4945]: I1008 15:46:00.043585 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5deefdab-a29a-4b5a-ad0d-08eccea748d1" path="/var/lib/kubelet/pods/5deefdab-a29a-4b5a-ad0d-08eccea748d1/volumes" Oct 08 15:46:00 crc kubenswrapper[4945]: I1008 15:46:00.044473 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d55dc803-9da2-4af4-b64a-6d9b40b2fa06" path="/var/lib/kubelet/pods/d55dc803-9da2-4af4-b64a-6d9b40b2fa06/volumes" Oct 08 15:46:37 crc kubenswrapper[4945]: I1008 15:46:37.363363 4945 scope.go:117] "RemoveContainer" containerID="8914d335e229a76e0b5acc4ba67784634f85d1084b9068152005b837fc496c79" Oct 08 15:46:37 crc kubenswrapper[4945]: I1008 15:46:37.404802 4945 scope.go:117] "RemoveContainer" containerID="de4fc6dd04fd18b220a8754fb329f78ade3462fabcbbeb2d90071fe1eea1787d" Oct 08 15:46:37 crc kubenswrapper[4945]: I1008 15:46:37.455930 4945 scope.go:117] "RemoveContainer" containerID="de1bf0ed5f75e054c3099e66e65dbc1dd77273c439f48b8cb404a009e1613a21" Oct 08 15:46:37 crc kubenswrapper[4945]: I1008 15:46:37.525209 4945 scope.go:117] "RemoveContainer" containerID="132a5d7c75f48d809e36d017e92f44ede7220c89ea8f80b926eff721625df2dc" Oct 08 15:46:37 crc kubenswrapper[4945]: I1008 15:46:37.584463 4945 scope.go:117] "RemoveContainer" containerID="811ed07885f90146efc50daa22a8f32f58b52c6d8d2443fb9766122e2be554c1" Oct 08 15:46:37 crc kubenswrapper[4945]: I1008 15:46:37.627527 4945 scope.go:117] "RemoveContainer" containerID="fbb92a6973d3c5dbb73613c7d37fc0e8876a26d3027f5d940253ed02c7adb22d" Oct 08 15:46:44 crc kubenswrapper[4945]: I1008 15:46:44.636269 4945 generic.go:334] "Generic (PLEG): container finished" podID="e59943cd-9572-4e6b-8a58-c0ec1283b2aa" containerID="14982dc341d12f6481e386bdbafc6f25c6978e93ccfe1e1dd5d68fcddbad447e" exitCode=0 Oct 08 15:46:44 crc kubenswrapper[4945]: I1008 15:46:44.636330 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" event={"ID":"e59943cd-9572-4e6b-8a58-c0ec1283b2aa","Type":"ContainerDied","Data":"14982dc341d12f6481e386bdbafc6f25c6978e93ccfe1e1dd5d68fcddbad447e"} Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.052806 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.186511 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4t5k\" (UniqueName: \"kubernetes.io/projected/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-kube-api-access-h4t5k\") pod \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.187032 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-ssh-key\") pod \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.187214 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-inventory\") pod \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\" (UID: \"e59943cd-9572-4e6b-8a58-c0ec1283b2aa\") " Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.193374 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-kube-api-access-h4t5k" (OuterVolumeSpecName: "kube-api-access-h4t5k") pod "e59943cd-9572-4e6b-8a58-c0ec1283b2aa" (UID: "e59943cd-9572-4e6b-8a58-c0ec1283b2aa"). InnerVolumeSpecName "kube-api-access-h4t5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.224379 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e59943cd-9572-4e6b-8a58-c0ec1283b2aa" (UID: "e59943cd-9572-4e6b-8a58-c0ec1283b2aa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.227248 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-inventory" (OuterVolumeSpecName: "inventory") pod "e59943cd-9572-4e6b-8a58-c0ec1283b2aa" (UID: "e59943cd-9572-4e6b-8a58-c0ec1283b2aa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.289593 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4t5k\" (UniqueName: \"kubernetes.io/projected/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-kube-api-access-h4t5k\") on node \"crc\" DevicePath \"\"" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.289623 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.289633 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59943cd-9572-4e6b-8a58-c0ec1283b2aa-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.660880 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" event={"ID":"e59943cd-9572-4e6b-8a58-c0ec1283b2aa","Type":"ContainerDied","Data":"81d74b6232e0f9bd61789440c6ee8498f4aace429d9eb70ee5ff754f5b02d1a4"} Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.661217 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81d74b6232e0f9bd61789440c6ee8498f4aace429d9eb70ee5ff754f5b02d1a4" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.660949 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.749435 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh"] Oct 08 15:46:46 crc kubenswrapper[4945]: E1008 15:46:46.749839 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7ea7eab-c495-4da0-bdaa-18edd47f0744" containerName="collect-profiles" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.750053 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7ea7eab-c495-4da0-bdaa-18edd47f0744" containerName="collect-profiles" Oct 08 15:46:46 crc kubenswrapper[4945]: E1008 15:46:46.750122 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59943cd-9572-4e6b-8a58-c0ec1283b2aa" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.750132 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59943cd-9572-4e6b-8a58-c0ec1283b2aa" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.750350 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7ea7eab-c495-4da0-bdaa-18edd47f0744" containerName="collect-profiles" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.750370 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e59943cd-9572-4e6b-8a58-c0ec1283b2aa" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.751043 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.753998 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.754526 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.754621 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.754524 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.765787 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh"] Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.799369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.799615 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj8kv\" (UniqueName: \"kubernetes.io/projected/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-kube-api-access-vj8kv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.799750 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.901767 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.902151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj8kv\" (UniqueName: \"kubernetes.io/projected/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-kube-api-access-vj8kv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.902269 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.912001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.914719 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:46 crc kubenswrapper[4945]: I1008 15:46:46.932283 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj8kv\" (UniqueName: \"kubernetes.io/projected/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-kube-api-access-vj8kv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:47 crc kubenswrapper[4945]: I1008 15:46:47.077857 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:46:47 crc kubenswrapper[4945]: I1008 15:46:47.609602 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh"] Oct 08 15:46:47 crc kubenswrapper[4945]: I1008 15:46:47.610128 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:46:47 crc kubenswrapper[4945]: I1008 15:46:47.670778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" event={"ID":"fcdb8897-6c75-48d4-bbdf-b4eb486331d5","Type":"ContainerStarted","Data":"f1ba6b4a30ca82d90a08d7ce1dbdcc6c2e9a04d3f0f6b631ec0b46ce2c230c60"} Oct 08 15:46:49 crc kubenswrapper[4945]: I1008 15:46:49.696723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" event={"ID":"fcdb8897-6c75-48d4-bbdf-b4eb486331d5","Type":"ContainerStarted","Data":"8fe8e9d82d6c99df6711efcd7d0ea0b97ead829042ee663d74e1b9d1a394c6d5"} Oct 08 15:46:49 crc kubenswrapper[4945]: I1008 15:46:49.723752 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" podStartSLOduration=2.6079050710000002 podStartE2EDuration="3.723730448s" podCreationTimestamp="2025-10-08 15:46:46 +0000 UTC" firstStartedPulling="2025-10-08 15:46:47.609912088 +0000 UTC m=+2016.963826989" lastFinishedPulling="2025-10-08 15:46:48.725737465 +0000 UTC m=+2018.079652366" observedRunningTime="2025-10-08 15:46:49.720250981 +0000 UTC m=+2019.074165902" watchObservedRunningTime="2025-10-08 15:46:49.723730448 +0000 UTC m=+2019.077645349" Oct 08 15:46:57 crc kubenswrapper[4945]: I1008 15:46:57.047939 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jh7mz"] Oct 08 15:46:57 crc kubenswrapper[4945]: I1008 15:46:57.059836 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jh7mz"] Oct 08 15:46:58 crc kubenswrapper[4945]: I1008 15:46:58.040789 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59d1ed3b-11c9-4deb-9947-605022b6592d" path="/var/lib/kubelet/pods/59d1ed3b-11c9-4deb-9947-605022b6592d/volumes" Oct 08 15:47:16 crc kubenswrapper[4945]: I1008 15:47:16.059852 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-tv6pp"] Oct 08 15:47:16 crc kubenswrapper[4945]: I1008 15:47:16.072848 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-tv6pp"] Oct 08 15:47:18 crc kubenswrapper[4945]: I1008 15:47:18.042727 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b2392b9-5374-4ee2-919d-545683412f06" path="/var/lib/kubelet/pods/6b2392b9-5374-4ee2-919d-545683412f06/volumes" Oct 08 15:47:19 crc kubenswrapper[4945]: I1008 15:47:19.184960 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:47:19 crc kubenswrapper[4945]: I1008 15:47:19.185056 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:47:21 crc kubenswrapper[4945]: I1008 15:47:21.036819 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zmh84"] Oct 08 15:47:21 crc kubenswrapper[4945]: I1008 15:47:21.044431 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zmh84"] Oct 08 15:47:22 crc kubenswrapper[4945]: I1008 15:47:22.037880 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb4085df-bd3d-4f38-b206-2cdadcf006bf" path="/var/lib/kubelet/pods/eb4085df-bd3d-4f38-b206-2cdadcf006bf/volumes" Oct 08 15:47:37 crc kubenswrapper[4945]: I1008 15:47:37.784007 4945 scope.go:117] "RemoveContainer" containerID="18655ccb6da9e928ef848ad30691e324ddbc0eabf0582d0dbad5cc1195ac3aa3" Oct 08 15:47:37 crc kubenswrapper[4945]: I1008 15:47:37.830956 4945 scope.go:117] "RemoveContainer" containerID="1bb696a9615feeb5c6ec70ad2cb044ebc1fa3b9e578521b98b35de44ac2be442" Oct 08 15:47:37 crc kubenswrapper[4945]: I1008 15:47:37.877402 4945 scope.go:117] "RemoveContainer" containerID="824d3c69503a0ec1cb20fc535ad482935e1ec3ef7f0790d2c4ade097a7d9399f" Oct 08 15:47:49 crc kubenswrapper[4945]: I1008 15:47:49.184655 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:47:49 crc kubenswrapper[4945]: I1008 15:47:49.185262 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:48:00 crc kubenswrapper[4945]: I1008 15:48:00.047043 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-xk6lb"] Oct 08 15:48:00 crc kubenswrapper[4945]: I1008 15:48:00.057656 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-xk6lb"] Oct 08 15:48:02 crc kubenswrapper[4945]: I1008 15:48:02.037313 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24aa0642-d709-4118-ad55-fd216646ecc2" path="/var/lib/kubelet/pods/24aa0642-d709-4118-ad55-fd216646ecc2/volumes" Oct 08 15:48:10 crc kubenswrapper[4945]: I1008 15:48:10.543878 4945 generic.go:334] "Generic (PLEG): container finished" podID="fcdb8897-6c75-48d4-bbdf-b4eb486331d5" containerID="8fe8e9d82d6c99df6711efcd7d0ea0b97ead829042ee663d74e1b9d1a394c6d5" exitCode=0 Oct 08 15:48:10 crc kubenswrapper[4945]: I1008 15:48:10.543969 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" event={"ID":"fcdb8897-6c75-48d4-bbdf-b4eb486331d5","Type":"ContainerDied","Data":"8fe8e9d82d6c99df6711efcd7d0ea0b97ead829042ee663d74e1b9d1a394c6d5"} Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.040211 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.168238 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-ssh-key\") pod \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.168412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj8kv\" (UniqueName: \"kubernetes.io/projected/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-kube-api-access-vj8kv\") pod \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.168470 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-inventory\") pod \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\" (UID: \"fcdb8897-6c75-48d4-bbdf-b4eb486331d5\") " Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.178459 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-kube-api-access-vj8kv" (OuterVolumeSpecName: "kube-api-access-vj8kv") pod "fcdb8897-6c75-48d4-bbdf-b4eb486331d5" (UID: "fcdb8897-6c75-48d4-bbdf-b4eb486331d5"). InnerVolumeSpecName "kube-api-access-vj8kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.216809 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-inventory" (OuterVolumeSpecName: "inventory") pod "fcdb8897-6c75-48d4-bbdf-b4eb486331d5" (UID: "fcdb8897-6c75-48d4-bbdf-b4eb486331d5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.237073 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fcdb8897-6c75-48d4-bbdf-b4eb486331d5" (UID: "fcdb8897-6c75-48d4-bbdf-b4eb486331d5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.273614 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj8kv\" (UniqueName: \"kubernetes.io/projected/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-kube-api-access-vj8kv\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.273658 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.273669 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcdb8897-6c75-48d4-bbdf-b4eb486331d5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.586758 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" event={"ID":"fcdb8897-6c75-48d4-bbdf-b4eb486331d5","Type":"ContainerDied","Data":"f1ba6b4a30ca82d90a08d7ce1dbdcc6c2e9a04d3f0f6b631ec0b46ce2c230c60"} Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.586809 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1ba6b4a30ca82d90a08d7ce1dbdcc6c2e9a04d3f0f6b631ec0b46ce2c230c60" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.586890 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.667567 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2"] Oct 08 15:48:12 crc kubenswrapper[4945]: E1008 15:48:12.668204 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcdb8897-6c75-48d4-bbdf-b4eb486331d5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.668225 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcdb8897-6c75-48d4-bbdf-b4eb486331d5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.668508 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcdb8897-6c75-48d4-bbdf-b4eb486331d5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.669439 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.671827 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.672405 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.672441 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.674648 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.677000 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2"] Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.783550 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4h5w\" (UniqueName: \"kubernetes.io/projected/53ec9f64-c34e-4765-b4ef-fd44c75f003e-kube-api-access-v4h5w\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.783651 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.784012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.886326 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4h5w\" (UniqueName: \"kubernetes.io/projected/53ec9f64-c34e-4765-b4ef-fd44c75f003e-kube-api-access-v4h5w\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.886416 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.886608 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.892672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.892791 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.910491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4h5w\" (UniqueName: \"kubernetes.io/projected/53ec9f64-c34e-4765-b4ef-fd44c75f003e-kube-api-access-v4h5w\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:12 crc kubenswrapper[4945]: I1008 15:48:12.985485 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:13 crc kubenswrapper[4945]: I1008 15:48:13.584747 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2"] Oct 08 15:48:14 crc kubenswrapper[4945]: I1008 15:48:14.609924 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" event={"ID":"53ec9f64-c34e-4765-b4ef-fd44c75f003e","Type":"ContainerStarted","Data":"f8ee9909b9335024c43a87e937057ed06ccb99d1c193242d91420062f0ae4c75"} Oct 08 15:48:14 crc kubenswrapper[4945]: I1008 15:48:14.610360 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" event={"ID":"53ec9f64-c34e-4765-b4ef-fd44c75f003e","Type":"ContainerStarted","Data":"5fb499b135e26a83ad3c81a2f4a2745aef30dcd8bb8a250548877743502e312a"} Oct 08 15:48:14 crc kubenswrapper[4945]: I1008 15:48:14.637631 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" podStartSLOduration=1.893334257 podStartE2EDuration="2.637612468s" podCreationTimestamp="2025-10-08 15:48:12 +0000 UTC" firstStartedPulling="2025-10-08 15:48:13.595016417 +0000 UTC m=+2102.948931318" lastFinishedPulling="2025-10-08 15:48:14.339294628 +0000 UTC m=+2103.693209529" observedRunningTime="2025-10-08 15:48:14.628651663 +0000 UTC m=+2103.982566564" watchObservedRunningTime="2025-10-08 15:48:14.637612468 +0000 UTC m=+2103.991527369" Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.184333 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.185133 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.185223 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.186297 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"00b2ab00906aa9fd3ef271df8c0b042025d08d3100676934af532ebbb72bb625"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.186389 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://00b2ab00906aa9fd3ef271df8c0b042025d08d3100676934af532ebbb72bb625" gracePeriod=600 Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.664025 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="00b2ab00906aa9fd3ef271df8c0b042025d08d3100676934af532ebbb72bb625" exitCode=0 Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.664106 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"00b2ab00906aa9fd3ef271df8c0b042025d08d3100676934af532ebbb72bb625"} Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.664400 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4"} Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.664418 4945 scope.go:117] "RemoveContainer" containerID="630c428ecb3116bd89a2a60780552ddc9ac3a3f672fa60345db472ccd132c2e0" Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.666236 4945 generic.go:334] "Generic (PLEG): container finished" podID="53ec9f64-c34e-4765-b4ef-fd44c75f003e" containerID="f8ee9909b9335024c43a87e937057ed06ccb99d1c193242d91420062f0ae4c75" exitCode=0 Oct 08 15:48:19 crc kubenswrapper[4945]: I1008 15:48:19.666280 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" event={"ID":"53ec9f64-c34e-4765-b4ef-fd44c75f003e","Type":"ContainerDied","Data":"f8ee9909b9335024c43a87e937057ed06ccb99d1c193242d91420062f0ae4c75"} Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.142789 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.179416 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-ssh-key\") pod \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.179510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4h5w\" (UniqueName: \"kubernetes.io/projected/53ec9f64-c34e-4765-b4ef-fd44c75f003e-kube-api-access-v4h5w\") pod \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.179535 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-inventory\") pod \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\" (UID: \"53ec9f64-c34e-4765-b4ef-fd44c75f003e\") " Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.185528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ec9f64-c34e-4765-b4ef-fd44c75f003e-kube-api-access-v4h5w" (OuterVolumeSpecName: "kube-api-access-v4h5w") pod "53ec9f64-c34e-4765-b4ef-fd44c75f003e" (UID: "53ec9f64-c34e-4765-b4ef-fd44c75f003e"). InnerVolumeSpecName "kube-api-access-v4h5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.216885 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-inventory" (OuterVolumeSpecName: "inventory") pod "53ec9f64-c34e-4765-b4ef-fd44c75f003e" (UID: "53ec9f64-c34e-4765-b4ef-fd44c75f003e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.217494 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53ec9f64-c34e-4765-b4ef-fd44c75f003e" (UID: "53ec9f64-c34e-4765-b4ef-fd44c75f003e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.282325 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.282363 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4h5w\" (UniqueName: \"kubernetes.io/projected/53ec9f64-c34e-4765-b4ef-fd44c75f003e-kube-api-access-v4h5w\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.282377 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53ec9f64-c34e-4765-b4ef-fd44c75f003e-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.700081 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" event={"ID":"53ec9f64-c34e-4765-b4ef-fd44c75f003e","Type":"ContainerDied","Data":"5fb499b135e26a83ad3c81a2f4a2745aef30dcd8bb8a250548877743502e312a"} Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.700403 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fb499b135e26a83ad3c81a2f4a2745aef30dcd8bb8a250548877743502e312a" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.700354 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.791929 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z"] Oct 08 15:48:21 crc kubenswrapper[4945]: E1008 15:48:21.792548 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ec9f64-c34e-4765-b4ef-fd44c75f003e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.792579 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ec9f64-c34e-4765-b4ef-fd44c75f003e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.792963 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ec9f64-c34e-4765-b4ef-fd44c75f003e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.796539 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.799210 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.799992 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.800438 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.801281 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.809001 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z"] Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.893799 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.996172 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.996255 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:21 crc kubenswrapper[4945]: I1008 15:48:21.996297 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gnmb\" (UniqueName: \"kubernetes.io/projected/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-kube-api-access-6gnmb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.000168 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.098639 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.099015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gnmb\" (UniqueName: \"kubernetes.io/projected/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-kube-api-access-6gnmb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.103649 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.116276 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gnmb\" (UniqueName: \"kubernetes.io/projected/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-kube-api-access-6gnmb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twd5z\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.118280 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.673351 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z"] Oct 08 15:48:22 crc kubenswrapper[4945]: W1008 15:48:22.680487 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbad79416_d5f9_4e87_bab7_cbd0ea8047e9.slice/crio-19f25838b84092fad747c075ad099409876b248de522b3a4250b02f351c71f91 WatchSource:0}: Error finding container 19f25838b84092fad747c075ad099409876b248de522b3a4250b02f351c71f91: Status 404 returned error can't find the container with id 19f25838b84092fad747c075ad099409876b248de522b3a4250b02f351c71f91 Oct 08 15:48:22 crc kubenswrapper[4945]: I1008 15:48:22.708809 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" event={"ID":"bad79416-d5f9-4e87-bab7-cbd0ea8047e9","Type":"ContainerStarted","Data":"19f25838b84092fad747c075ad099409876b248de522b3a4250b02f351c71f91"} Oct 08 15:48:23 crc kubenswrapper[4945]: I1008 15:48:23.719838 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" event={"ID":"bad79416-d5f9-4e87-bab7-cbd0ea8047e9","Type":"ContainerStarted","Data":"28a5682cbdde1e0683e7d825c9976f1d4a34313e04c2ed92b24c17bb9946685d"} Oct 08 15:48:23 crc kubenswrapper[4945]: I1008 15:48:23.743461 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" podStartSLOduration=2.211409785 podStartE2EDuration="2.743436105s" podCreationTimestamp="2025-10-08 15:48:21 +0000 UTC" firstStartedPulling="2025-10-08 15:48:22.68436103 +0000 UTC m=+2112.038275951" lastFinishedPulling="2025-10-08 15:48:23.21638736 +0000 UTC m=+2112.570302271" observedRunningTime="2025-10-08 15:48:23.738656745 +0000 UTC m=+2113.092571656" watchObservedRunningTime="2025-10-08 15:48:23.743436105 +0000 UTC m=+2113.097351016" Oct 08 15:48:37 crc kubenswrapper[4945]: I1008 15:48:37.999022 4945 scope.go:117] "RemoveContainer" containerID="f0eacbb9d4443ec00ff22d374cb725dbcce021dfc4d84790858a190abf0810d0" Oct 08 15:49:01 crc kubenswrapper[4945]: I1008 15:49:01.099211 4945 generic.go:334] "Generic (PLEG): container finished" podID="bad79416-d5f9-4e87-bab7-cbd0ea8047e9" containerID="28a5682cbdde1e0683e7d825c9976f1d4a34313e04c2ed92b24c17bb9946685d" exitCode=0 Oct 08 15:49:01 crc kubenswrapper[4945]: I1008 15:49:01.099368 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" event={"ID":"bad79416-d5f9-4e87-bab7-cbd0ea8047e9","Type":"ContainerDied","Data":"28a5682cbdde1e0683e7d825c9976f1d4a34313e04c2ed92b24c17bb9946685d"} Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.525888 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.708888 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-ssh-key\") pod \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.709216 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-inventory\") pod \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.709262 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gnmb\" (UniqueName: \"kubernetes.io/projected/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-kube-api-access-6gnmb\") pod \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\" (UID: \"bad79416-d5f9-4e87-bab7-cbd0ea8047e9\") " Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.714794 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-kube-api-access-6gnmb" (OuterVolumeSpecName: "kube-api-access-6gnmb") pod "bad79416-d5f9-4e87-bab7-cbd0ea8047e9" (UID: "bad79416-d5f9-4e87-bab7-cbd0ea8047e9"). InnerVolumeSpecName "kube-api-access-6gnmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.737352 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bad79416-d5f9-4e87-bab7-cbd0ea8047e9" (UID: "bad79416-d5f9-4e87-bab7-cbd0ea8047e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.748269 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-inventory" (OuterVolumeSpecName: "inventory") pod "bad79416-d5f9-4e87-bab7-cbd0ea8047e9" (UID: "bad79416-d5f9-4e87-bab7-cbd0ea8047e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.811821 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.812026 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gnmb\" (UniqueName: \"kubernetes.io/projected/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-kube-api-access-6gnmb\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:02 crc kubenswrapper[4945]: I1008 15:49:02.812170 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bad79416-d5f9-4e87-bab7-cbd0ea8047e9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.122994 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" event={"ID":"bad79416-d5f9-4e87-bab7-cbd0ea8047e9","Type":"ContainerDied","Data":"19f25838b84092fad747c075ad099409876b248de522b3a4250b02f351c71f91"} Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.123069 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19f25838b84092fad747c075ad099409876b248de522b3a4250b02f351c71f91" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.123223 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twd5z" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.214831 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps"] Oct 08 15:49:03 crc kubenswrapper[4945]: E1008 15:49:03.215766 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad79416-d5f9-4e87-bab7-cbd0ea8047e9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.215789 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad79416-d5f9-4e87-bab7-cbd0ea8047e9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.215987 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad79416-d5f9-4e87-bab7-cbd0ea8047e9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.216743 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.224442 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.224618 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.224670 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.224769 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.235177 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps"] Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.320750 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.320854 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.321074 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtpn6\" (UniqueName: \"kubernetes.io/projected/71b8f45e-f47d-432d-b22e-3fa39e8059dd-kube-api-access-qtpn6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.423633 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.423799 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.423961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtpn6\" (UniqueName: \"kubernetes.io/projected/71b8f45e-f47d-432d-b22e-3fa39e8059dd-kube-api-access-qtpn6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.428934 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.429156 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.440665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtpn6\" (UniqueName: \"kubernetes.io/projected/71b8f45e-f47d-432d-b22e-3fa39e8059dd-kube-api-access-qtpn6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tkrps\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:03 crc kubenswrapper[4945]: I1008 15:49:03.543768 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:49:04 crc kubenswrapper[4945]: I1008 15:49:04.110154 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps"] Oct 08 15:49:04 crc kubenswrapper[4945]: W1008 15:49:04.116610 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71b8f45e_f47d_432d_b22e_3fa39e8059dd.slice/crio-9459db05b231fa5c9092419a34575e495198a8d2bf141e4bcb1eb737b58905c6 WatchSource:0}: Error finding container 9459db05b231fa5c9092419a34575e495198a8d2bf141e4bcb1eb737b58905c6: Status 404 returned error can't find the container with id 9459db05b231fa5c9092419a34575e495198a8d2bf141e4bcb1eb737b58905c6 Oct 08 15:49:04 crc kubenswrapper[4945]: I1008 15:49:04.133137 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" event={"ID":"71b8f45e-f47d-432d-b22e-3fa39e8059dd","Type":"ContainerStarted","Data":"9459db05b231fa5c9092419a34575e495198a8d2bf141e4bcb1eb737b58905c6"} Oct 08 15:49:06 crc kubenswrapper[4945]: I1008 15:49:06.151982 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" event={"ID":"71b8f45e-f47d-432d-b22e-3fa39e8059dd","Type":"ContainerStarted","Data":"29f00a58982564ef19c9f10c501724a41d0d832c8d03fb579bbbb28d98cce52e"} Oct 08 15:49:06 crc kubenswrapper[4945]: I1008 15:49:06.173490 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" podStartSLOduration=1.945542558 podStartE2EDuration="3.173472856s" podCreationTimestamp="2025-10-08 15:49:03 +0000 UTC" firstStartedPulling="2025-10-08 15:49:04.118315007 +0000 UTC m=+2153.472229908" lastFinishedPulling="2025-10-08 15:49:05.346245285 +0000 UTC m=+2154.700160206" observedRunningTime="2025-10-08 15:49:06.169076086 +0000 UTC m=+2155.522990987" watchObservedRunningTime="2025-10-08 15:49:06.173472856 +0000 UTC m=+2155.527387757" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.629771 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-flzr6"] Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.633428 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.659941 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flzr6"] Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.760086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z84tn\" (UniqueName: \"kubernetes.io/projected/8bae0907-1b77-446a-8c77-f7dffe76b545-kube-api-access-z84tn\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.760330 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-catalog-content\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.760374 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-utilities\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.862383 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-catalog-content\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.862434 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-utilities\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.862479 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z84tn\" (UniqueName: \"kubernetes.io/projected/8bae0907-1b77-446a-8c77-f7dffe76b545-kube-api-access-z84tn\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.862990 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-catalog-content\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.863027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-utilities\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.885653 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z84tn\" (UniqueName: \"kubernetes.io/projected/8bae0907-1b77-446a-8c77-f7dffe76b545-kube-api-access-z84tn\") pod \"community-operators-flzr6\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:09 crc kubenswrapper[4945]: I1008 15:49:09.955847 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:10 crc kubenswrapper[4945]: I1008 15:49:10.478966 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flzr6"] Oct 08 15:49:10 crc kubenswrapper[4945]: W1008 15:49:10.479700 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bae0907_1b77_446a_8c77_f7dffe76b545.slice/crio-254dc529f7f6104839d6689d768d52713edd89ca134bfccb3e23073caf3ae9c2 WatchSource:0}: Error finding container 254dc529f7f6104839d6689d768d52713edd89ca134bfccb3e23073caf3ae9c2: Status 404 returned error can't find the container with id 254dc529f7f6104839d6689d768d52713edd89ca134bfccb3e23073caf3ae9c2 Oct 08 15:49:11 crc kubenswrapper[4945]: I1008 15:49:11.197792 4945 generic.go:334] "Generic (PLEG): container finished" podID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerID="31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15" exitCode=0 Oct 08 15:49:11 crc kubenswrapper[4945]: I1008 15:49:11.197878 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flzr6" event={"ID":"8bae0907-1b77-446a-8c77-f7dffe76b545","Type":"ContainerDied","Data":"31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15"} Oct 08 15:49:11 crc kubenswrapper[4945]: I1008 15:49:11.198396 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flzr6" event={"ID":"8bae0907-1b77-446a-8c77-f7dffe76b545","Type":"ContainerStarted","Data":"254dc529f7f6104839d6689d768d52713edd89ca134bfccb3e23073caf3ae9c2"} Oct 08 15:49:13 crc kubenswrapper[4945]: I1008 15:49:13.215861 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flzr6" event={"ID":"8bae0907-1b77-446a-8c77-f7dffe76b545","Type":"ContainerStarted","Data":"d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67"} Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.304249 4945 generic.go:334] "Generic (PLEG): container finished" podID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerID="d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67" exitCode=0 Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.304315 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flzr6" event={"ID":"8bae0907-1b77-446a-8c77-f7dffe76b545","Type":"ContainerDied","Data":"d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67"} Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.755394 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fnsch"] Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.758753 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.767420 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnsch"] Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.914064 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-catalog-content\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.914121 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl5kz\" (UniqueName: \"kubernetes.io/projected/85584174-b1c7-423a-8cdd-2b608b773cdf-kube-api-access-hl5kz\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:21 crc kubenswrapper[4945]: I1008 15:49:21.914276 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-utilities\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.016428 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-utilities\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.016537 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-catalog-content\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.016585 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl5kz\" (UniqueName: \"kubernetes.io/projected/85584174-b1c7-423a-8cdd-2b608b773cdf-kube-api-access-hl5kz\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.016937 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-utilities\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.017097 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-catalog-content\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.036914 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl5kz\" (UniqueName: \"kubernetes.io/projected/85584174-b1c7-423a-8cdd-2b608b773cdf-kube-api-access-hl5kz\") pod \"redhat-marketplace-fnsch\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.080742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:22 crc kubenswrapper[4945]: I1008 15:49:22.585973 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnsch"] Oct 08 15:49:22 crc kubenswrapper[4945]: W1008 15:49:22.589226 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85584174_b1c7_423a_8cdd_2b608b773cdf.slice/crio-f1129e7b82ce76c690552cf3038c001739dc533fd99404e3253b1272a5d93f70 WatchSource:0}: Error finding container f1129e7b82ce76c690552cf3038c001739dc533fd99404e3253b1272a5d93f70: Status 404 returned error can't find the container with id f1129e7b82ce76c690552cf3038c001739dc533fd99404e3253b1272a5d93f70 Oct 08 15:49:23 crc kubenswrapper[4945]: I1008 15:49:23.333702 4945 generic.go:334] "Generic (PLEG): container finished" podID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerID="358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877" exitCode=0 Oct 08 15:49:23 crc kubenswrapper[4945]: I1008 15:49:23.333793 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnsch" event={"ID":"85584174-b1c7-423a-8cdd-2b608b773cdf","Type":"ContainerDied","Data":"358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877"} Oct 08 15:49:23 crc kubenswrapper[4945]: I1008 15:49:23.334069 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnsch" event={"ID":"85584174-b1c7-423a-8cdd-2b608b773cdf","Type":"ContainerStarted","Data":"f1129e7b82ce76c690552cf3038c001739dc533fd99404e3253b1272a5d93f70"} Oct 08 15:49:23 crc kubenswrapper[4945]: I1008 15:49:23.337516 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flzr6" event={"ID":"8bae0907-1b77-446a-8c77-f7dffe76b545","Type":"ContainerStarted","Data":"0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31"} Oct 08 15:49:23 crc kubenswrapper[4945]: I1008 15:49:23.384243 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-flzr6" podStartSLOduration=3.4177523 podStartE2EDuration="14.384218025s" podCreationTimestamp="2025-10-08 15:49:09 +0000 UTC" firstStartedPulling="2025-10-08 15:49:11.19970865 +0000 UTC m=+2160.553623551" lastFinishedPulling="2025-10-08 15:49:22.166174375 +0000 UTC m=+2171.520089276" observedRunningTime="2025-10-08 15:49:23.376025099 +0000 UTC m=+2172.729940000" watchObservedRunningTime="2025-10-08 15:49:23.384218025 +0000 UTC m=+2172.738132956" Oct 08 15:49:25 crc kubenswrapper[4945]: I1008 15:49:25.357185 4945 generic.go:334] "Generic (PLEG): container finished" podID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerID="b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a" exitCode=0 Oct 08 15:49:25 crc kubenswrapper[4945]: I1008 15:49:25.357364 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnsch" event={"ID":"85584174-b1c7-423a-8cdd-2b608b773cdf","Type":"ContainerDied","Data":"b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a"} Oct 08 15:49:26 crc kubenswrapper[4945]: I1008 15:49:26.390757 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnsch" event={"ID":"85584174-b1c7-423a-8cdd-2b608b773cdf","Type":"ContainerStarted","Data":"3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605"} Oct 08 15:49:26 crc kubenswrapper[4945]: I1008 15:49:26.412011 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fnsch" podStartSLOduration=2.938486662 podStartE2EDuration="5.41199065s" podCreationTimestamp="2025-10-08 15:49:21 +0000 UTC" firstStartedPulling="2025-10-08 15:49:23.336175521 +0000 UTC m=+2172.690090422" lastFinishedPulling="2025-10-08 15:49:25.809679509 +0000 UTC m=+2175.163594410" observedRunningTime="2025-10-08 15:49:26.408532314 +0000 UTC m=+2175.762447235" watchObservedRunningTime="2025-10-08 15:49:26.41199065 +0000 UTC m=+2175.765905551" Oct 08 15:49:29 crc kubenswrapper[4945]: I1008 15:49:29.956425 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:29 crc kubenswrapper[4945]: I1008 15:49:29.957032 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:30 crc kubenswrapper[4945]: I1008 15:49:30.000811 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:30 crc kubenswrapper[4945]: I1008 15:49:30.482231 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:30 crc kubenswrapper[4945]: I1008 15:49:30.528087 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flzr6"] Oct 08 15:49:32 crc kubenswrapper[4945]: I1008 15:49:32.081911 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:32 crc kubenswrapper[4945]: I1008 15:49:32.082498 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:32 crc kubenswrapper[4945]: I1008 15:49:32.127463 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:32 crc kubenswrapper[4945]: I1008 15:49:32.441764 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-flzr6" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="registry-server" containerID="cri-o://0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31" gracePeriod=2 Oct 08 15:49:32 crc kubenswrapper[4945]: I1008 15:49:32.504853 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:32 crc kubenswrapper[4945]: I1008 15:49:32.924267 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.040488 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-utilities\") pod \"8bae0907-1b77-446a-8c77-f7dffe76b545\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.040552 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z84tn\" (UniqueName: \"kubernetes.io/projected/8bae0907-1b77-446a-8c77-f7dffe76b545-kube-api-access-z84tn\") pod \"8bae0907-1b77-446a-8c77-f7dffe76b545\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.040833 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-catalog-content\") pod \"8bae0907-1b77-446a-8c77-f7dffe76b545\" (UID: \"8bae0907-1b77-446a-8c77-f7dffe76b545\") " Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.041191 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-utilities" (OuterVolumeSpecName: "utilities") pod "8bae0907-1b77-446a-8c77-f7dffe76b545" (UID: "8bae0907-1b77-446a-8c77-f7dffe76b545"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.041546 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.046335 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bae0907-1b77-446a-8c77-f7dffe76b545-kube-api-access-z84tn" (OuterVolumeSpecName: "kube-api-access-z84tn") pod "8bae0907-1b77-446a-8c77-f7dffe76b545" (UID: "8bae0907-1b77-446a-8c77-f7dffe76b545"). InnerVolumeSpecName "kube-api-access-z84tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.101828 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bae0907-1b77-446a-8c77-f7dffe76b545" (UID: "8bae0907-1b77-446a-8c77-f7dffe76b545"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.144216 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bae0907-1b77-446a-8c77-f7dffe76b545-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.144266 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z84tn\" (UniqueName: \"kubernetes.io/projected/8bae0907-1b77-446a-8c77-f7dffe76b545-kube-api-access-z84tn\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.458054 4945 generic.go:334] "Generic (PLEG): container finished" podID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerID="0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31" exitCode=0 Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.458158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flzr6" event={"ID":"8bae0907-1b77-446a-8c77-f7dffe76b545","Type":"ContainerDied","Data":"0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31"} Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.458195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flzr6" event={"ID":"8bae0907-1b77-446a-8c77-f7dffe76b545","Type":"ContainerDied","Data":"254dc529f7f6104839d6689d768d52713edd89ca134bfccb3e23073caf3ae9c2"} Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.458201 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flzr6" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.458252 4945 scope.go:117] "RemoveContainer" containerID="0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.486607 4945 scope.go:117] "RemoveContainer" containerID="d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.504292 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flzr6"] Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.512568 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-flzr6"] Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.526790 4945 scope.go:117] "RemoveContainer" containerID="31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.552214 4945 scope.go:117] "RemoveContainer" containerID="0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31" Oct 08 15:49:33 crc kubenswrapper[4945]: E1008 15:49:33.552863 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31\": container with ID starting with 0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31 not found: ID does not exist" containerID="0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.552889 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31"} err="failed to get container status \"0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31\": rpc error: code = NotFound desc = could not find container \"0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31\": container with ID starting with 0bf369e90fd175d8dc53e846b54eb817c03c3bb7bfc6319f5466472a47016a31 not found: ID does not exist" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.552909 4945 scope.go:117] "RemoveContainer" containerID="d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67" Oct 08 15:49:33 crc kubenswrapper[4945]: E1008 15:49:33.553150 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67\": container with ID starting with d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67 not found: ID does not exist" containerID="d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.553170 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67"} err="failed to get container status \"d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67\": rpc error: code = NotFound desc = could not find container \"d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67\": container with ID starting with d537b7b45dc275555ebb44d3f8b443a0f9aafe3c19db8f7446b7cee4872b5d67 not found: ID does not exist" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.553187 4945 scope.go:117] "RemoveContainer" containerID="31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15" Oct 08 15:49:33 crc kubenswrapper[4945]: E1008 15:49:33.553495 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15\": container with ID starting with 31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15 not found: ID does not exist" containerID="31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15" Oct 08 15:49:33 crc kubenswrapper[4945]: I1008 15:49:33.553585 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15"} err="failed to get container status \"31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15\": rpc error: code = NotFound desc = could not find container \"31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15\": container with ID starting with 31e5765cad8b64824d87610ccd71b87a06f5829b163ff4eed073712f1637ed15 not found: ID does not exist" Oct 08 15:49:34 crc kubenswrapper[4945]: I1008 15:49:34.036769 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" path="/var/lib/kubelet/pods/8bae0907-1b77-446a-8c77-f7dffe76b545/volumes" Oct 08 15:49:34 crc kubenswrapper[4945]: I1008 15:49:34.235029 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnsch"] Oct 08 15:49:35 crc kubenswrapper[4945]: I1008 15:49:35.485628 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fnsch" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="registry-server" containerID="cri-o://3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605" gracePeriod=2 Oct 08 15:49:35 crc kubenswrapper[4945]: I1008 15:49:35.941559 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:35 crc kubenswrapper[4945]: I1008 15:49:35.997401 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-catalog-content\") pod \"85584174-b1c7-423a-8cdd-2b608b773cdf\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " Oct 08 15:49:35 crc kubenswrapper[4945]: I1008 15:49:35.997446 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl5kz\" (UniqueName: \"kubernetes.io/projected/85584174-b1c7-423a-8cdd-2b608b773cdf-kube-api-access-hl5kz\") pod \"85584174-b1c7-423a-8cdd-2b608b773cdf\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " Oct 08 15:49:35 crc kubenswrapper[4945]: I1008 15:49:35.997466 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-utilities\") pod \"85584174-b1c7-423a-8cdd-2b608b773cdf\" (UID: \"85584174-b1c7-423a-8cdd-2b608b773cdf\") " Oct 08 15:49:35 crc kubenswrapper[4945]: I1008 15:49:35.999594 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-utilities" (OuterVolumeSpecName: "utilities") pod "85584174-b1c7-423a-8cdd-2b608b773cdf" (UID: "85584174-b1c7-423a-8cdd-2b608b773cdf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.003060 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85584174-b1c7-423a-8cdd-2b608b773cdf-kube-api-access-hl5kz" (OuterVolumeSpecName: "kube-api-access-hl5kz") pod "85584174-b1c7-423a-8cdd-2b608b773cdf" (UID: "85584174-b1c7-423a-8cdd-2b608b773cdf"). InnerVolumeSpecName "kube-api-access-hl5kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.013061 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85584174-b1c7-423a-8cdd-2b608b773cdf" (UID: "85584174-b1c7-423a-8cdd-2b608b773cdf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.099887 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl5kz\" (UniqueName: \"kubernetes.io/projected/85584174-b1c7-423a-8cdd-2b608b773cdf-kube-api-access-hl5kz\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.099926 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.099935 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85584174-b1c7-423a-8cdd-2b608b773cdf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.497648 4945 generic.go:334] "Generic (PLEG): container finished" podID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerID="3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605" exitCode=0 Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.497691 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnsch" event={"ID":"85584174-b1c7-423a-8cdd-2b608b773cdf","Type":"ContainerDied","Data":"3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605"} Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.497722 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnsch" event={"ID":"85584174-b1c7-423a-8cdd-2b608b773cdf","Type":"ContainerDied","Data":"f1129e7b82ce76c690552cf3038c001739dc533fd99404e3253b1272a5d93f70"} Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.497731 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnsch" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.497766 4945 scope.go:117] "RemoveContainer" containerID="3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.524799 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnsch"] Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.534080 4945 scope.go:117] "RemoveContainer" containerID="b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.537918 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnsch"] Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.562327 4945 scope.go:117] "RemoveContainer" containerID="358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.617952 4945 scope.go:117] "RemoveContainer" containerID="3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605" Oct 08 15:49:36 crc kubenswrapper[4945]: E1008 15:49:36.618515 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605\": container with ID starting with 3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605 not found: ID does not exist" containerID="3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.618560 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605"} err="failed to get container status \"3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605\": rpc error: code = NotFound desc = could not find container \"3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605\": container with ID starting with 3c3ed6673bd947a6571c9d726983ddc06c9c2391f6e0a32adc89af0e38753605 not found: ID does not exist" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.618587 4945 scope.go:117] "RemoveContainer" containerID="b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a" Oct 08 15:49:36 crc kubenswrapper[4945]: E1008 15:49:36.618910 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a\": container with ID starting with b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a not found: ID does not exist" containerID="b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.618937 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a"} err="failed to get container status \"b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a\": rpc error: code = NotFound desc = could not find container \"b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a\": container with ID starting with b47169333187f6b38e86a93fd342a59fcf165dcadfa4c9fcc590afb11515cb6a not found: ID does not exist" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.618951 4945 scope.go:117] "RemoveContainer" containerID="358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877" Oct 08 15:49:36 crc kubenswrapper[4945]: E1008 15:49:36.619207 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877\": container with ID starting with 358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877 not found: ID does not exist" containerID="358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877" Oct 08 15:49:36 crc kubenswrapper[4945]: I1008 15:49:36.619238 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877"} err="failed to get container status \"358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877\": rpc error: code = NotFound desc = could not find container \"358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877\": container with ID starting with 358748261b6812a8ff137d3b9828b7d20f0ae1861f3c66b577c49533f6a9d877 not found: ID does not exist" Oct 08 15:49:38 crc kubenswrapper[4945]: I1008 15:49:38.036293 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" path="/var/lib/kubelet/pods/85584174-b1c7-423a-8cdd-2b608b773cdf/volumes" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.932330 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vbcwt"] Oct 08 15:49:59 crc kubenswrapper[4945]: E1008 15:49:59.933342 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="extract-content" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933363 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="extract-content" Oct 08 15:49:59 crc kubenswrapper[4945]: E1008 15:49:59.933396 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="registry-server" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933406 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="registry-server" Oct 08 15:49:59 crc kubenswrapper[4945]: E1008 15:49:59.933424 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="extract-content" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933433 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="extract-content" Oct 08 15:49:59 crc kubenswrapper[4945]: E1008 15:49:59.933449 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="extract-utilities" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933465 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="extract-utilities" Oct 08 15:49:59 crc kubenswrapper[4945]: E1008 15:49:59.933476 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="extract-utilities" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933483 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="extract-utilities" Oct 08 15:49:59 crc kubenswrapper[4945]: E1008 15:49:59.933493 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="registry-server" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933502 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="registry-server" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933811 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="85584174-b1c7-423a-8cdd-2b608b773cdf" containerName="registry-server" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.933829 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bae0907-1b77-446a-8c77-f7dffe76b545" containerName="registry-server" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.936255 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:49:59 crc kubenswrapper[4945]: I1008 15:49:59.951271 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbcwt"] Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.123070 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-catalog-content\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.123422 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pzbd\" (UniqueName: \"kubernetes.io/projected/300a362b-6e2a-4497-bca8-00983aa8b5f2-kube-api-access-8pzbd\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.123578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-utilities\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.225845 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-utilities\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.226012 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-catalog-content\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.226058 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pzbd\" (UniqueName: \"kubernetes.io/projected/300a362b-6e2a-4497-bca8-00983aa8b5f2-kube-api-access-8pzbd\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.226412 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-utilities\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.226474 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-catalog-content\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.249052 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pzbd\" (UniqueName: \"kubernetes.io/projected/300a362b-6e2a-4497-bca8-00983aa8b5f2-kube-api-access-8pzbd\") pod \"certified-operators-vbcwt\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.260016 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:00 crc kubenswrapper[4945]: I1008 15:50:00.802812 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbcwt"] Oct 08 15:50:01 crc kubenswrapper[4945]: I1008 15:50:01.738305 4945 generic.go:334] "Generic (PLEG): container finished" podID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerID="09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342" exitCode=0 Oct 08 15:50:01 crc kubenswrapper[4945]: I1008 15:50:01.738365 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbcwt" event={"ID":"300a362b-6e2a-4497-bca8-00983aa8b5f2","Type":"ContainerDied","Data":"09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342"} Oct 08 15:50:01 crc kubenswrapper[4945]: I1008 15:50:01.738626 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbcwt" event={"ID":"300a362b-6e2a-4497-bca8-00983aa8b5f2","Type":"ContainerStarted","Data":"ed6d2e387cdc395626f287f25862ae70f6fd772048b86944eb1e98b22af52af4"} Oct 08 15:50:02 crc kubenswrapper[4945]: I1008 15:50:02.748407 4945 generic.go:334] "Generic (PLEG): container finished" podID="71b8f45e-f47d-432d-b22e-3fa39e8059dd" containerID="29f00a58982564ef19c9f10c501724a41d0d832c8d03fb579bbbb28d98cce52e" exitCode=2 Oct 08 15:50:02 crc kubenswrapper[4945]: I1008 15:50:02.748512 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" event={"ID":"71b8f45e-f47d-432d-b22e-3fa39e8059dd","Type":"ContainerDied","Data":"29f00a58982564ef19c9f10c501724a41d0d832c8d03fb579bbbb28d98cce52e"} Oct 08 15:50:02 crc kubenswrapper[4945]: I1008 15:50:02.752260 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbcwt" event={"ID":"300a362b-6e2a-4497-bca8-00983aa8b5f2","Type":"ContainerStarted","Data":"46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1"} Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.166656 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.318016 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtpn6\" (UniqueName: \"kubernetes.io/projected/71b8f45e-f47d-432d-b22e-3fa39e8059dd-kube-api-access-qtpn6\") pod \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.318338 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-ssh-key\") pod \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.318383 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-inventory\") pod \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\" (UID: \"71b8f45e-f47d-432d-b22e-3fa39e8059dd\") " Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.326323 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b8f45e-f47d-432d-b22e-3fa39e8059dd-kube-api-access-qtpn6" (OuterVolumeSpecName: "kube-api-access-qtpn6") pod "71b8f45e-f47d-432d-b22e-3fa39e8059dd" (UID: "71b8f45e-f47d-432d-b22e-3fa39e8059dd"). InnerVolumeSpecName "kube-api-access-qtpn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.350604 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-inventory" (OuterVolumeSpecName: "inventory") pod "71b8f45e-f47d-432d-b22e-3fa39e8059dd" (UID: "71b8f45e-f47d-432d-b22e-3fa39e8059dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.352557 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "71b8f45e-f47d-432d-b22e-3fa39e8059dd" (UID: "71b8f45e-f47d-432d-b22e-3fa39e8059dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.420978 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.421013 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtpn6\" (UniqueName: \"kubernetes.io/projected/71b8f45e-f47d-432d-b22e-3fa39e8059dd-kube-api-access-qtpn6\") on node \"crc\" DevicePath \"\"" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.421028 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b8f45e-f47d-432d-b22e-3fa39e8059dd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.783761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" event={"ID":"71b8f45e-f47d-432d-b22e-3fa39e8059dd","Type":"ContainerDied","Data":"9459db05b231fa5c9092419a34575e495198a8d2bf141e4bcb1eb737b58905c6"} Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.783811 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9459db05b231fa5c9092419a34575e495198a8d2bf141e4bcb1eb737b58905c6" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.783849 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tkrps" Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.787929 4945 generic.go:334] "Generic (PLEG): container finished" podID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerID="46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1" exitCode=0 Oct 08 15:50:04 crc kubenswrapper[4945]: I1008 15:50:04.787986 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbcwt" event={"ID":"300a362b-6e2a-4497-bca8-00983aa8b5f2","Type":"ContainerDied","Data":"46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1"} Oct 08 15:50:05 crc kubenswrapper[4945]: I1008 15:50:05.799570 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbcwt" event={"ID":"300a362b-6e2a-4497-bca8-00983aa8b5f2","Type":"ContainerStarted","Data":"5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388"} Oct 08 15:50:05 crc kubenswrapper[4945]: I1008 15:50:05.825923 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vbcwt" podStartSLOduration=3.292977783 podStartE2EDuration="6.825900325s" podCreationTimestamp="2025-10-08 15:49:59 +0000 UTC" firstStartedPulling="2025-10-08 15:50:01.741910947 +0000 UTC m=+2211.095825848" lastFinishedPulling="2025-10-08 15:50:05.274833489 +0000 UTC m=+2214.628748390" observedRunningTime="2025-10-08 15:50:05.820934801 +0000 UTC m=+2215.174849722" watchObservedRunningTime="2025-10-08 15:50:05.825900325 +0000 UTC m=+2215.179815236" Oct 08 15:50:10 crc kubenswrapper[4945]: I1008 15:50:10.260541 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:10 crc kubenswrapper[4945]: I1008 15:50:10.261221 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:10 crc kubenswrapper[4945]: I1008 15:50:10.311823 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:10 crc kubenswrapper[4945]: I1008 15:50:10.914273 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:10 crc kubenswrapper[4945]: I1008 15:50:10.964975 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbcwt"] Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.048861 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t"] Oct 08 15:50:12 crc kubenswrapper[4945]: E1008 15:50:12.049587 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b8f45e-f47d-432d-b22e-3fa39e8059dd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.049599 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b8f45e-f47d-432d-b22e-3fa39e8059dd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.049792 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b8f45e-f47d-432d-b22e-3fa39e8059dd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.050638 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.054256 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.054450 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.056801 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.056862 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.069912 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t"] Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.084373 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcc42\" (UniqueName: \"kubernetes.io/projected/a6323567-dc30-45ed-8611-4925d983e0ef-kube-api-access-zcc42\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.084474 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.084547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.186748 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcc42\" (UniqueName: \"kubernetes.io/projected/a6323567-dc30-45ed-8611-4925d983e0ef-kube-api-access-zcc42\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.186795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.186826 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.199590 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.201665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.202393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcc42\" (UniqueName: \"kubernetes.io/projected/a6323567-dc30-45ed-8611-4925d983e0ef-kube-api-access-zcc42\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.391079 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.890706 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vbcwt" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="registry-server" containerID="cri-o://5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388" gracePeriod=2 Oct 08 15:50:12 crc kubenswrapper[4945]: I1008 15:50:12.967257 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t"] Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.366769 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.512277 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pzbd\" (UniqueName: \"kubernetes.io/projected/300a362b-6e2a-4497-bca8-00983aa8b5f2-kube-api-access-8pzbd\") pod \"300a362b-6e2a-4497-bca8-00983aa8b5f2\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.512355 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-catalog-content\") pod \"300a362b-6e2a-4497-bca8-00983aa8b5f2\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.512480 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-utilities\") pod \"300a362b-6e2a-4497-bca8-00983aa8b5f2\" (UID: \"300a362b-6e2a-4497-bca8-00983aa8b5f2\") " Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.517589 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-utilities" (OuterVolumeSpecName: "utilities") pod "300a362b-6e2a-4497-bca8-00983aa8b5f2" (UID: "300a362b-6e2a-4497-bca8-00983aa8b5f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.520951 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/300a362b-6e2a-4497-bca8-00983aa8b5f2-kube-api-access-8pzbd" (OuterVolumeSpecName: "kube-api-access-8pzbd") pod "300a362b-6e2a-4497-bca8-00983aa8b5f2" (UID: "300a362b-6e2a-4497-bca8-00983aa8b5f2"). InnerVolumeSpecName "kube-api-access-8pzbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.557577 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "300a362b-6e2a-4497-bca8-00983aa8b5f2" (UID: "300a362b-6e2a-4497-bca8-00983aa8b5f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.615078 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pzbd\" (UniqueName: \"kubernetes.io/projected/300a362b-6e2a-4497-bca8-00983aa8b5f2-kube-api-access-8pzbd\") on node \"crc\" DevicePath \"\"" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.615669 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.615752 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300a362b-6e2a-4497-bca8-00983aa8b5f2-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.905721 4945 generic.go:334] "Generic (PLEG): container finished" podID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerID="5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388" exitCode=0 Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.905799 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbcwt" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.905792 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbcwt" event={"ID":"300a362b-6e2a-4497-bca8-00983aa8b5f2","Type":"ContainerDied","Data":"5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388"} Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.905870 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbcwt" event={"ID":"300a362b-6e2a-4497-bca8-00983aa8b5f2","Type":"ContainerDied","Data":"ed6d2e387cdc395626f287f25862ae70f6fd772048b86944eb1e98b22af52af4"} Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.905895 4945 scope.go:117] "RemoveContainer" containerID="5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.908150 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" event={"ID":"a6323567-dc30-45ed-8611-4925d983e0ef","Type":"ContainerStarted","Data":"f11ed5347ba28eb1c55c31daf740c8a57870888b709ca86a4274c2afd1838664"} Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.942921 4945 scope.go:117] "RemoveContainer" containerID="46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.950185 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbcwt"] Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.958478 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vbcwt"] Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.972950 4945 scope.go:117] "RemoveContainer" containerID="09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.996788 4945 scope.go:117] "RemoveContainer" containerID="5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388" Oct 08 15:50:13 crc kubenswrapper[4945]: E1008 15:50:13.997357 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388\": container with ID starting with 5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388 not found: ID does not exist" containerID="5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.997399 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388"} err="failed to get container status \"5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388\": rpc error: code = NotFound desc = could not find container \"5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388\": container with ID starting with 5d72acbd23bba85967210a6eb4a3b4a8e9aff522b3e5e29d8c9ec774d8e77388 not found: ID does not exist" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.997426 4945 scope.go:117] "RemoveContainer" containerID="46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1" Oct 08 15:50:13 crc kubenswrapper[4945]: E1008 15:50:13.997720 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1\": container with ID starting with 46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1 not found: ID does not exist" containerID="46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.997749 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1"} err="failed to get container status \"46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1\": rpc error: code = NotFound desc = could not find container \"46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1\": container with ID starting with 46b62d9bae184c94101d174db68a21c06ae7d053c6695264b0405c6337dae4e1 not found: ID does not exist" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.997767 4945 scope.go:117] "RemoveContainer" containerID="09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342" Oct 08 15:50:13 crc kubenswrapper[4945]: E1008 15:50:13.998189 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342\": container with ID starting with 09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342 not found: ID does not exist" containerID="09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342" Oct 08 15:50:13 crc kubenswrapper[4945]: I1008 15:50:13.998236 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342"} err="failed to get container status \"09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342\": rpc error: code = NotFound desc = could not find container \"09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342\": container with ID starting with 09601352aa8fe8bdc4d99471d9e1b26c75a0af14e41f8cd383400a10aea5c342 not found: ID does not exist" Oct 08 15:50:14 crc kubenswrapper[4945]: I1008 15:50:14.036089 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" path="/var/lib/kubelet/pods/300a362b-6e2a-4497-bca8-00983aa8b5f2/volumes" Oct 08 15:50:14 crc kubenswrapper[4945]: I1008 15:50:14.920036 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" event={"ID":"a6323567-dc30-45ed-8611-4925d983e0ef","Type":"ContainerStarted","Data":"aee22f8daab9e402858b5792e43c861244f1f6fe4c9d9d98f81d0508b9548165"} Oct 08 15:50:14 crc kubenswrapper[4945]: I1008 15:50:14.956558 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" podStartSLOduration=2.131483694 podStartE2EDuration="2.95652192s" podCreationTimestamp="2025-10-08 15:50:12 +0000 UTC" firstStartedPulling="2025-10-08 15:50:12.994122886 +0000 UTC m=+2222.348037777" lastFinishedPulling="2025-10-08 15:50:13.819161102 +0000 UTC m=+2223.173076003" observedRunningTime="2025-10-08 15:50:14.94257589 +0000 UTC m=+2224.296490801" watchObservedRunningTime="2025-10-08 15:50:14.95652192 +0000 UTC m=+2224.310436871" Oct 08 15:50:19 crc kubenswrapper[4945]: I1008 15:50:19.184476 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:50:19 crc kubenswrapper[4945]: I1008 15:50:19.185025 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:50:49 crc kubenswrapper[4945]: I1008 15:50:49.183927 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:50:49 crc kubenswrapper[4945]: I1008 15:50:49.184641 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.256494 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b8f87"] Oct 08 15:50:53 crc kubenswrapper[4945]: E1008 15:50:53.258217 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="extract-utilities" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.258242 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="extract-utilities" Oct 08 15:50:53 crc kubenswrapper[4945]: E1008 15:50:53.258391 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="registry-server" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.258408 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="registry-server" Oct 08 15:50:53 crc kubenswrapper[4945]: E1008 15:50:53.258460 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="extract-content" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.258474 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="extract-content" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.259343 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="300a362b-6e2a-4497-bca8-00983aa8b5f2" containerName="registry-server" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.264442 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.290473 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b8f87"] Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.327152 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-catalog-content\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.327296 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm2jw\" (UniqueName: \"kubernetes.io/projected/c70e801d-3f47-4d6d-9974-44ad602e3f52-kube-api-access-rm2jw\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.327331 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-utilities\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.429782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm2jw\" (UniqueName: \"kubernetes.io/projected/c70e801d-3f47-4d6d-9974-44ad602e3f52-kube-api-access-rm2jw\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.429983 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-utilities\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.430064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-catalog-content\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.430627 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-catalog-content\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.430778 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-utilities\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.449338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm2jw\" (UniqueName: \"kubernetes.io/projected/c70e801d-3f47-4d6d-9974-44ad602e3f52-kube-api-access-rm2jw\") pod \"redhat-operators-b8f87\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:53 crc kubenswrapper[4945]: I1008 15:50:53.595237 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:50:54 crc kubenswrapper[4945]: I1008 15:50:54.092777 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b8f87"] Oct 08 15:50:54 crc kubenswrapper[4945]: W1008 15:50:54.098916 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc70e801d_3f47_4d6d_9974_44ad602e3f52.slice/crio-a4d1d5564f81a71ef7d0ade61d65582f528c2089b3043b8fafa887712e86ccd9 WatchSource:0}: Error finding container a4d1d5564f81a71ef7d0ade61d65582f528c2089b3043b8fafa887712e86ccd9: Status 404 returned error can't find the container with id a4d1d5564f81a71ef7d0ade61d65582f528c2089b3043b8fafa887712e86ccd9 Oct 08 15:50:54 crc kubenswrapper[4945]: I1008 15:50:54.325086 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8f87" event={"ID":"c70e801d-3f47-4d6d-9974-44ad602e3f52","Type":"ContainerStarted","Data":"a4d1d5564f81a71ef7d0ade61d65582f528c2089b3043b8fafa887712e86ccd9"} Oct 08 15:50:55 crc kubenswrapper[4945]: I1008 15:50:55.335017 4945 generic.go:334] "Generic (PLEG): container finished" podID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerID="e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757" exitCode=0 Oct 08 15:50:55 crc kubenswrapper[4945]: I1008 15:50:55.335094 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8f87" event={"ID":"c70e801d-3f47-4d6d-9974-44ad602e3f52","Type":"ContainerDied","Data":"e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757"} Oct 08 15:50:58 crc kubenswrapper[4945]: I1008 15:50:58.365090 4945 generic.go:334] "Generic (PLEG): container finished" podID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerID="ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3" exitCode=0 Oct 08 15:50:58 crc kubenswrapper[4945]: I1008 15:50:58.365227 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8f87" event={"ID":"c70e801d-3f47-4d6d-9974-44ad602e3f52","Type":"ContainerDied","Data":"ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3"} Oct 08 15:51:01 crc kubenswrapper[4945]: I1008 15:51:01.395173 4945 generic.go:334] "Generic (PLEG): container finished" podID="a6323567-dc30-45ed-8611-4925d983e0ef" containerID="aee22f8daab9e402858b5792e43c861244f1f6fe4c9d9d98f81d0508b9548165" exitCode=0 Oct 08 15:51:01 crc kubenswrapper[4945]: I1008 15:51:01.395378 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" event={"ID":"a6323567-dc30-45ed-8611-4925d983e0ef","Type":"ContainerDied","Data":"aee22f8daab9e402858b5792e43c861244f1f6fe4c9d9d98f81d0508b9548165"} Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.406437 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8f87" event={"ID":"c70e801d-3f47-4d6d-9974-44ad602e3f52","Type":"ContainerStarted","Data":"5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede"} Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.436980 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b8f87" podStartSLOduration=3.534262231 podStartE2EDuration="9.436958131s" podCreationTimestamp="2025-10-08 15:50:53 +0000 UTC" firstStartedPulling="2025-10-08 15:50:55.33734991 +0000 UTC m=+2264.691264801" lastFinishedPulling="2025-10-08 15:51:01.24004578 +0000 UTC m=+2270.593960701" observedRunningTime="2025-10-08 15:51:02.429096933 +0000 UTC m=+2271.783011834" watchObservedRunningTime="2025-10-08 15:51:02.436958131 +0000 UTC m=+2271.790873042" Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.827135 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.922256 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcc42\" (UniqueName: \"kubernetes.io/projected/a6323567-dc30-45ed-8611-4925d983e0ef-kube-api-access-zcc42\") pod \"a6323567-dc30-45ed-8611-4925d983e0ef\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.922370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-ssh-key\") pod \"a6323567-dc30-45ed-8611-4925d983e0ef\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.922548 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-inventory\") pod \"a6323567-dc30-45ed-8611-4925d983e0ef\" (UID: \"a6323567-dc30-45ed-8611-4925d983e0ef\") " Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.929371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6323567-dc30-45ed-8611-4925d983e0ef-kube-api-access-zcc42" (OuterVolumeSpecName: "kube-api-access-zcc42") pod "a6323567-dc30-45ed-8611-4925d983e0ef" (UID: "a6323567-dc30-45ed-8611-4925d983e0ef"). InnerVolumeSpecName "kube-api-access-zcc42". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.953761 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-inventory" (OuterVolumeSpecName: "inventory") pod "a6323567-dc30-45ed-8611-4925d983e0ef" (UID: "a6323567-dc30-45ed-8611-4925d983e0ef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:51:02 crc kubenswrapper[4945]: I1008 15:51:02.956680 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a6323567-dc30-45ed-8611-4925d983e0ef" (UID: "a6323567-dc30-45ed-8611-4925d983e0ef"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.024621 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcc42\" (UniqueName: \"kubernetes.io/projected/a6323567-dc30-45ed-8611-4925d983e0ef-kube-api-access-zcc42\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.025026 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.025314 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6323567-dc30-45ed-8611-4925d983e0ef-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.416463 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.416469 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t" event={"ID":"a6323567-dc30-45ed-8611-4925d983e0ef","Type":"ContainerDied","Data":"f11ed5347ba28eb1c55c31daf740c8a57870888b709ca86a4274c2afd1838664"} Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.416531 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f11ed5347ba28eb1c55c31daf740c8a57870888b709ca86a4274c2afd1838664" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.544218 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bzn58"] Oct 08 15:51:03 crc kubenswrapper[4945]: E1008 15:51:03.544718 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6323567-dc30-45ed-8611-4925d983e0ef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.544742 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6323567-dc30-45ed-8611-4925d983e0ef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.552396 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6323567-dc30-45ed-8611-4925d983e0ef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.553371 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.560724 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.560764 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.560996 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.561215 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.597351 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.597427 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.611176 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bzn58"] Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.751648 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.751939 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.751993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhm6k\" (UniqueName: \"kubernetes.io/projected/02a8cc7d-9ec0-4840-9023-f2550c396fcc-kube-api-access-dhm6k\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.853934 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.854338 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.854400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhm6k\" (UniqueName: \"kubernetes.io/projected/02a8cc7d-9ec0-4840-9023-f2550c396fcc-kube-api-access-dhm6k\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.860398 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.861967 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.874157 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhm6k\" (UniqueName: \"kubernetes.io/projected/02a8cc7d-9ec0-4840-9023-f2550c396fcc-kube-api-access-dhm6k\") pod \"ssh-known-hosts-edpm-deployment-bzn58\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:03 crc kubenswrapper[4945]: I1008 15:51:03.881451 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:04 crc kubenswrapper[4945]: I1008 15:51:04.476885 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bzn58"] Oct 08 15:51:04 crc kubenswrapper[4945]: I1008 15:51:04.706751 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b8f87" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="registry-server" probeResult="failure" output=< Oct 08 15:51:04 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 15:51:04 crc kubenswrapper[4945]: > Oct 08 15:51:05 crc kubenswrapper[4945]: I1008 15:51:05.434840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" event={"ID":"02a8cc7d-9ec0-4840-9023-f2550c396fcc","Type":"ContainerStarted","Data":"6a90b29e98ac74997bc8f51583174bcdaebee368914ae7ec4e6cdbad6081cae7"} Oct 08 15:51:06 crc kubenswrapper[4945]: I1008 15:51:06.446260 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" event={"ID":"02a8cc7d-9ec0-4840-9023-f2550c396fcc","Type":"ContainerStarted","Data":"33a4840ff9f88b3ba1baabcbc4749b96d18364b748b2dc4f9e0e81357f9bf7fd"} Oct 08 15:51:06 crc kubenswrapper[4945]: I1008 15:51:06.462947 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" podStartSLOduration=2.739781844 podStartE2EDuration="3.462930045s" podCreationTimestamp="2025-10-08 15:51:03 +0000 UTC" firstStartedPulling="2025-10-08 15:51:04.485447073 +0000 UTC m=+2273.839361984" lastFinishedPulling="2025-10-08 15:51:05.208595284 +0000 UTC m=+2274.562510185" observedRunningTime="2025-10-08 15:51:06.4595568 +0000 UTC m=+2275.813471701" watchObservedRunningTime="2025-10-08 15:51:06.462930045 +0000 UTC m=+2275.816844946" Oct 08 15:51:13 crc kubenswrapper[4945]: I1008 15:51:13.523507 4945 generic.go:334] "Generic (PLEG): container finished" podID="02a8cc7d-9ec0-4840-9023-f2550c396fcc" containerID="33a4840ff9f88b3ba1baabcbc4749b96d18364b748b2dc4f9e0e81357f9bf7fd" exitCode=0 Oct 08 15:51:13 crc kubenswrapper[4945]: I1008 15:51:13.523659 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" event={"ID":"02a8cc7d-9ec0-4840-9023-f2550c396fcc","Type":"ContainerDied","Data":"33a4840ff9f88b3ba1baabcbc4749b96d18364b748b2dc4f9e0e81357f9bf7fd"} Oct 08 15:51:13 crc kubenswrapper[4945]: I1008 15:51:13.672637 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:51:13 crc kubenswrapper[4945]: I1008 15:51:13.746366 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:51:13 crc kubenswrapper[4945]: I1008 15:51:13.921765 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b8f87"] Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.041860 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.189333 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-ssh-key-openstack-edpm-ipam\") pod \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.189484 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhm6k\" (UniqueName: \"kubernetes.io/projected/02a8cc7d-9ec0-4840-9023-f2550c396fcc-kube-api-access-dhm6k\") pod \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.189594 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-inventory-0\") pod \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\" (UID: \"02a8cc7d-9ec0-4840-9023-f2550c396fcc\") " Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.197313 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a8cc7d-9ec0-4840-9023-f2550c396fcc-kube-api-access-dhm6k" (OuterVolumeSpecName: "kube-api-access-dhm6k") pod "02a8cc7d-9ec0-4840-9023-f2550c396fcc" (UID: "02a8cc7d-9ec0-4840-9023-f2550c396fcc"). InnerVolumeSpecName "kube-api-access-dhm6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.220710 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "02a8cc7d-9ec0-4840-9023-f2550c396fcc" (UID: "02a8cc7d-9ec0-4840-9023-f2550c396fcc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.231300 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "02a8cc7d-9ec0-4840-9023-f2550c396fcc" (UID: "02a8cc7d-9ec0-4840-9023-f2550c396fcc"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.292066 4945 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.292130 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/02a8cc7d-9ec0-4840-9023-f2550c396fcc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.292147 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhm6k\" (UniqueName: \"kubernetes.io/projected/02a8cc7d-9ec0-4840-9023-f2550c396fcc-kube-api-access-dhm6k\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.553913 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" event={"ID":"02a8cc7d-9ec0-4840-9023-f2550c396fcc","Type":"ContainerDied","Data":"6a90b29e98ac74997bc8f51583174bcdaebee368914ae7ec4e6cdbad6081cae7"} Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.554416 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a90b29e98ac74997bc8f51583174bcdaebee368914ae7ec4e6cdbad6081cae7" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.553994 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bzn58" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.553962 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b8f87" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="registry-server" containerID="cri-o://5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede" gracePeriod=2 Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.638275 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn"] Oct 08 15:51:15 crc kubenswrapper[4945]: E1008 15:51:15.639216 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02a8cc7d-9ec0-4840-9023-f2550c396fcc" containerName="ssh-known-hosts-edpm-deployment" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.639239 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="02a8cc7d-9ec0-4840-9023-f2550c396fcc" containerName="ssh-known-hosts-edpm-deployment" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.639534 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="02a8cc7d-9ec0-4840-9023-f2550c396fcc" containerName="ssh-known-hosts-edpm-deployment" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.641023 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.646601 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.646761 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.646896 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.646979 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.652127 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn"] Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.804352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.804516 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.804565 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvz6m\" (UniqueName: \"kubernetes.io/projected/3a5d1441-c9ee-46f1-8004-361e72448f64-kube-api-access-rvz6m\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.906318 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.906388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.906428 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvz6m\" (UniqueName: \"kubernetes.io/projected/3a5d1441-c9ee-46f1-8004-361e72448f64-kube-api-access-rvz6m\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.913616 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.914568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:15 crc kubenswrapper[4945]: I1008 15:51:15.925735 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvz6m\" (UniqueName: \"kubernetes.io/projected/3a5d1441-c9ee-46f1-8004-361e72448f64-kube-api-access-rvz6m\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfrnn\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.004996 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.027144 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.111714 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-utilities\") pod \"c70e801d-3f47-4d6d-9974-44ad602e3f52\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.111871 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm2jw\" (UniqueName: \"kubernetes.io/projected/c70e801d-3f47-4d6d-9974-44ad602e3f52-kube-api-access-rm2jw\") pod \"c70e801d-3f47-4d6d-9974-44ad602e3f52\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.112809 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-catalog-content\") pod \"c70e801d-3f47-4d6d-9974-44ad602e3f52\" (UID: \"c70e801d-3f47-4d6d-9974-44ad602e3f52\") " Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.112837 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-utilities" (OuterVolumeSpecName: "utilities") pod "c70e801d-3f47-4d6d-9974-44ad602e3f52" (UID: "c70e801d-3f47-4d6d-9974-44ad602e3f52"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.113605 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.120613 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c70e801d-3f47-4d6d-9974-44ad602e3f52-kube-api-access-rm2jw" (OuterVolumeSpecName: "kube-api-access-rm2jw") pod "c70e801d-3f47-4d6d-9974-44ad602e3f52" (UID: "c70e801d-3f47-4d6d-9974-44ad602e3f52"). InnerVolumeSpecName "kube-api-access-rm2jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.204251 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c70e801d-3f47-4d6d-9974-44ad602e3f52" (UID: "c70e801d-3f47-4d6d-9974-44ad602e3f52"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.215844 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70e801d-3f47-4d6d-9974-44ad602e3f52-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.215880 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm2jw\" (UniqueName: \"kubernetes.io/projected/c70e801d-3f47-4d6d-9974-44ad602e3f52-kube-api-access-rm2jw\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:16 crc kubenswrapper[4945]: W1008 15:51:16.558010 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a5d1441_c9ee_46f1_8004_361e72448f64.slice/crio-e7ff6192f2fc22f9035b63a4cbbb139698509b4f95d91f6671d86df90c41eca9 WatchSource:0}: Error finding container e7ff6192f2fc22f9035b63a4cbbb139698509b4f95d91f6671d86df90c41eca9: Status 404 returned error can't find the container with id e7ff6192f2fc22f9035b63a4cbbb139698509b4f95d91f6671d86df90c41eca9 Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.558585 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn"] Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.569978 4945 generic.go:334] "Generic (PLEG): container finished" podID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerID="5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede" exitCode=0 Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.570025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8f87" event={"ID":"c70e801d-3f47-4d6d-9974-44ad602e3f52","Type":"ContainerDied","Data":"5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede"} Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.570062 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8f87" event={"ID":"c70e801d-3f47-4d6d-9974-44ad602e3f52","Type":"ContainerDied","Data":"a4d1d5564f81a71ef7d0ade61d65582f528c2089b3043b8fafa887712e86ccd9"} Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.570077 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8f87" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.570090 4945 scope.go:117] "RemoveContainer" containerID="5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.628614 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b8f87"] Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.629494 4945 scope.go:117] "RemoveContainer" containerID="ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.636919 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b8f87"] Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.650746 4945 scope.go:117] "RemoveContainer" containerID="e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.668979 4945 scope.go:117] "RemoveContainer" containerID="5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede" Oct 08 15:51:16 crc kubenswrapper[4945]: E1008 15:51:16.669414 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede\": container with ID starting with 5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede not found: ID does not exist" containerID="5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.669473 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede"} err="failed to get container status \"5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede\": rpc error: code = NotFound desc = could not find container \"5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede\": container with ID starting with 5dd999c8e4f2fcae3b600d9e9068e7e56ab3c11c2f2bf9e1976c0793db586ede not found: ID does not exist" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.669504 4945 scope.go:117] "RemoveContainer" containerID="ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3" Oct 08 15:51:16 crc kubenswrapper[4945]: E1008 15:51:16.669862 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3\": container with ID starting with ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3 not found: ID does not exist" containerID="ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.669892 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3"} err="failed to get container status \"ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3\": rpc error: code = NotFound desc = could not find container \"ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3\": container with ID starting with ddd475239d451910d73269094be67fb8fc5904ae13d280509f3e5276032f87d3 not found: ID does not exist" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.669916 4945 scope.go:117] "RemoveContainer" containerID="e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757" Oct 08 15:51:16 crc kubenswrapper[4945]: E1008 15:51:16.670162 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757\": container with ID starting with e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757 not found: ID does not exist" containerID="e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757" Oct 08 15:51:16 crc kubenswrapper[4945]: I1008 15:51:16.670192 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757"} err="failed to get container status \"e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757\": rpc error: code = NotFound desc = could not find container \"e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757\": container with ID starting with e75dfec09e6a82b5c01a94a46ea1484e3199137fffacaabfd4109e4dd669f757 not found: ID does not exist" Oct 08 15:51:17 crc kubenswrapper[4945]: I1008 15:51:17.579772 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" event={"ID":"3a5d1441-c9ee-46f1-8004-361e72448f64","Type":"ContainerStarted","Data":"e7ff6192f2fc22f9035b63a4cbbb139698509b4f95d91f6671d86df90c41eca9"} Oct 08 15:51:18 crc kubenswrapper[4945]: I1008 15:51:18.047724 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" path="/var/lib/kubelet/pods/c70e801d-3f47-4d6d-9974-44ad602e3f52/volumes" Oct 08 15:51:18 crc kubenswrapper[4945]: I1008 15:51:18.589070 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" event={"ID":"3a5d1441-c9ee-46f1-8004-361e72448f64","Type":"ContainerStarted","Data":"bdfc6f211f057469df9f0b0a9a0b6275886423784f0713c56e577cb04b38f9d9"} Oct 08 15:51:18 crc kubenswrapper[4945]: I1008 15:51:18.618199 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" podStartSLOduration=2.277383878 podStartE2EDuration="3.618181466s" podCreationTimestamp="2025-10-08 15:51:15 +0000 UTC" firstStartedPulling="2025-10-08 15:51:16.561088988 +0000 UTC m=+2285.915003889" lastFinishedPulling="2025-10-08 15:51:17.901886536 +0000 UTC m=+2287.255801477" observedRunningTime="2025-10-08 15:51:18.60716999 +0000 UTC m=+2287.961084891" watchObservedRunningTime="2025-10-08 15:51:18.618181466 +0000 UTC m=+2287.972096367" Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.183676 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.184170 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.184300 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.185329 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.185472 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" gracePeriod=600 Oct 08 15:51:19 crc kubenswrapper[4945]: E1008 15:51:19.318479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:51:19 crc kubenswrapper[4945]: E1008 15:51:19.348037 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode39fc950_7f86_4d6b_b4b0_25daafb66250.slice/crio-conmon-6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode39fc950_7f86_4d6b_b4b0_25daafb66250.slice/crio-6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4.scope\": RecentStats: unable to find data in memory cache]" Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.606166 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" exitCode=0 Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.606272 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4"} Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.606362 4945 scope.go:117] "RemoveContainer" containerID="00b2ab00906aa9fd3ef271df8c0b042025d08d3100676934af532ebbb72bb625" Oct 08 15:51:19 crc kubenswrapper[4945]: I1008 15:51:19.607704 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:51:19 crc kubenswrapper[4945]: E1008 15:51:19.608419 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:51:26 crc kubenswrapper[4945]: I1008 15:51:26.689778 4945 generic.go:334] "Generic (PLEG): container finished" podID="3a5d1441-c9ee-46f1-8004-361e72448f64" containerID="bdfc6f211f057469df9f0b0a9a0b6275886423784f0713c56e577cb04b38f9d9" exitCode=0 Oct 08 15:51:26 crc kubenswrapper[4945]: I1008 15:51:26.689863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" event={"ID":"3a5d1441-c9ee-46f1-8004-361e72448f64","Type":"ContainerDied","Data":"bdfc6f211f057469df9f0b0a9a0b6275886423784f0713c56e577cb04b38f9d9"} Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.130525 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.288187 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-ssh-key\") pod \"3a5d1441-c9ee-46f1-8004-361e72448f64\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.288425 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvz6m\" (UniqueName: \"kubernetes.io/projected/3a5d1441-c9ee-46f1-8004-361e72448f64-kube-api-access-rvz6m\") pod \"3a5d1441-c9ee-46f1-8004-361e72448f64\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.288461 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-inventory\") pod \"3a5d1441-c9ee-46f1-8004-361e72448f64\" (UID: \"3a5d1441-c9ee-46f1-8004-361e72448f64\") " Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.295565 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a5d1441-c9ee-46f1-8004-361e72448f64-kube-api-access-rvz6m" (OuterVolumeSpecName: "kube-api-access-rvz6m") pod "3a5d1441-c9ee-46f1-8004-361e72448f64" (UID: "3a5d1441-c9ee-46f1-8004-361e72448f64"). InnerVolumeSpecName "kube-api-access-rvz6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.316762 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-inventory" (OuterVolumeSpecName: "inventory") pod "3a5d1441-c9ee-46f1-8004-361e72448f64" (UID: "3a5d1441-c9ee-46f1-8004-361e72448f64"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.337287 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a5d1441-c9ee-46f1-8004-361e72448f64" (UID: "3a5d1441-c9ee-46f1-8004-361e72448f64"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.391877 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.392438 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvz6m\" (UniqueName: \"kubernetes.io/projected/3a5d1441-c9ee-46f1-8004-361e72448f64-kube-api-access-rvz6m\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.392685 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a5d1441-c9ee-46f1-8004-361e72448f64-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.712521 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" event={"ID":"3a5d1441-c9ee-46f1-8004-361e72448f64","Type":"ContainerDied","Data":"e7ff6192f2fc22f9035b63a4cbbb139698509b4f95d91f6671d86df90c41eca9"} Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.712930 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7ff6192f2fc22f9035b63a4cbbb139698509b4f95d91f6671d86df90c41eca9" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.712569 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfrnn" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.803547 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k"] Oct 08 15:51:28 crc kubenswrapper[4945]: E1008 15:51:28.803956 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="extract-utilities" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.803972 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="extract-utilities" Oct 08 15:51:28 crc kubenswrapper[4945]: E1008 15:51:28.803995 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="registry-server" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.804001 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="registry-server" Oct 08 15:51:28 crc kubenswrapper[4945]: E1008 15:51:28.804020 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="extract-content" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.804029 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="extract-content" Oct 08 15:51:28 crc kubenswrapper[4945]: E1008 15:51:28.804052 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5d1441-c9ee-46f1-8004-361e72448f64" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.804059 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5d1441-c9ee-46f1-8004-361e72448f64" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.804252 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70e801d-3f47-4d6d-9974-44ad602e3f52" containerName="registry-server" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.804274 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a5d1441-c9ee-46f1-8004-361e72448f64" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.805084 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.807288 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.807799 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.807906 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.808021 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.819600 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k"] Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.903924 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98fgl\" (UniqueName: \"kubernetes.io/projected/47e40524-5bcd-4e95-9353-c17f7c23476f-kube-api-access-98fgl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.904249 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:28 crc kubenswrapper[4945]: I1008 15:51:28.904415 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.006580 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.006733 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98fgl\" (UniqueName: \"kubernetes.io/projected/47e40524-5bcd-4e95-9353-c17f7c23476f-kube-api-access-98fgl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.006770 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.011521 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.011683 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.037994 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98fgl\" (UniqueName: \"kubernetes.io/projected/47e40524-5bcd-4e95-9353-c17f7c23476f-kube-api-access-98fgl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.127766 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.664963 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k"] Oct 08 15:51:29 crc kubenswrapper[4945]: I1008 15:51:29.733221 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" event={"ID":"47e40524-5bcd-4e95-9353-c17f7c23476f","Type":"ContainerStarted","Data":"0f61d421895820d0a8e04962d7c93fb808f6f1766058068c4fb8bf82a029bb73"} Oct 08 15:51:30 crc kubenswrapper[4945]: I1008 15:51:30.746496 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" event={"ID":"47e40524-5bcd-4e95-9353-c17f7c23476f","Type":"ContainerStarted","Data":"766c455a9d82fe2f844338a872d314637784b97d4952f40ebd8a1afe373df492"} Oct 08 15:51:30 crc kubenswrapper[4945]: I1008 15:51:30.778334 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" podStartSLOduration=2.198704607 podStartE2EDuration="2.778295929s" podCreationTimestamp="2025-10-08 15:51:28 +0000 UTC" firstStartedPulling="2025-10-08 15:51:29.67515275 +0000 UTC m=+2299.029067661" lastFinishedPulling="2025-10-08 15:51:30.254744092 +0000 UTC m=+2299.608658983" observedRunningTime="2025-10-08 15:51:30.768311708 +0000 UTC m=+2300.122226669" watchObservedRunningTime="2025-10-08 15:51:30.778295929 +0000 UTC m=+2300.132210870" Oct 08 15:51:34 crc kubenswrapper[4945]: I1008 15:51:34.025056 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:51:34 crc kubenswrapper[4945]: E1008 15:51:34.026081 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:51:49 crc kubenswrapper[4945]: I1008 15:51:49.024447 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:51:49 crc kubenswrapper[4945]: E1008 15:51:49.026197 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:52:00 crc kubenswrapper[4945]: I1008 15:52:00.024290 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:52:00 crc kubenswrapper[4945]: E1008 15:52:00.025082 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:52:14 crc kubenswrapper[4945]: I1008 15:52:14.024442 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:52:14 crc kubenswrapper[4945]: E1008 15:52:14.025549 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:52:25 crc kubenswrapper[4945]: I1008 15:52:25.024006 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:52:25 crc kubenswrapper[4945]: E1008 15:52:25.024686 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:52:37 crc kubenswrapper[4945]: I1008 15:52:37.024453 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:52:37 crc kubenswrapper[4945]: E1008 15:52:37.025233 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:52:46 crc kubenswrapper[4945]: I1008 15:52:46.593719 4945 generic.go:334] "Generic (PLEG): container finished" podID="47e40524-5bcd-4e95-9353-c17f7c23476f" containerID="766c455a9d82fe2f844338a872d314637784b97d4952f40ebd8a1afe373df492" exitCode=0 Oct 08 15:52:46 crc kubenswrapper[4945]: I1008 15:52:46.593848 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" event={"ID":"47e40524-5bcd-4e95-9353-c17f7c23476f","Type":"ContainerDied","Data":"766c455a9d82fe2f844338a872d314637784b97d4952f40ebd8a1afe373df492"} Oct 08 15:52:47 crc kubenswrapper[4945]: I1008 15:52:47.980281 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.003860 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-ssh-key\") pod \"47e40524-5bcd-4e95-9353-c17f7c23476f\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.004008 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-inventory\") pod \"47e40524-5bcd-4e95-9353-c17f7c23476f\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.004059 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98fgl\" (UniqueName: \"kubernetes.io/projected/47e40524-5bcd-4e95-9353-c17f7c23476f-kube-api-access-98fgl\") pod \"47e40524-5bcd-4e95-9353-c17f7c23476f\" (UID: \"47e40524-5bcd-4e95-9353-c17f7c23476f\") " Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.015462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47e40524-5bcd-4e95-9353-c17f7c23476f-kube-api-access-98fgl" (OuterVolumeSpecName: "kube-api-access-98fgl") pod "47e40524-5bcd-4e95-9353-c17f7c23476f" (UID: "47e40524-5bcd-4e95-9353-c17f7c23476f"). InnerVolumeSpecName "kube-api-access-98fgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.042038 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-inventory" (OuterVolumeSpecName: "inventory") pod "47e40524-5bcd-4e95-9353-c17f7c23476f" (UID: "47e40524-5bcd-4e95-9353-c17f7c23476f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.055190 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "47e40524-5bcd-4e95-9353-c17f7c23476f" (UID: "47e40524-5bcd-4e95-9353-c17f7c23476f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.106732 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.106778 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98fgl\" (UniqueName: \"kubernetes.io/projected/47e40524-5bcd-4e95-9353-c17f7c23476f-kube-api-access-98fgl\") on node \"crc\" DevicePath \"\"" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.106791 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47e40524-5bcd-4e95-9353-c17f7c23476f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.618370 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" event={"ID":"47e40524-5bcd-4e95-9353-c17f7c23476f","Type":"ContainerDied","Data":"0f61d421895820d0a8e04962d7c93fb808f6f1766058068c4fb8bf82a029bb73"} Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.618732 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f61d421895820d0a8e04962d7c93fb808f6f1766058068c4fb8bf82a029bb73" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.618451 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.715688 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d"] Oct 08 15:52:48 crc kubenswrapper[4945]: E1008 15:52:48.719450 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e40524-5bcd-4e95-9353-c17f7c23476f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.719497 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e40524-5bcd-4e95-9353-c17f7c23476f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.719881 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e40524-5bcd-4e95-9353-c17f7c23476f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.720788 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.724890 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.727479 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.727917 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.728058 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.728142 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.728333 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.728451 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.728617 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.730033 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d"] Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823428 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823500 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qqhl\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-kube-api-access-8qqhl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823593 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823627 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823654 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823730 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823758 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823793 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823907 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.823986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.824046 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.824074 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.824098 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.926536 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.926714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.926836 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.926901 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.926947 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927015 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927188 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qqhl\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-kube-api-access-8qqhl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927325 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927459 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927500 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.927554 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.932493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.933077 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.934053 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.934393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.934817 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.935284 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.935372 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.935491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.935314 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.935564 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.935735 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.936191 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.938022 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:48 crc kubenswrapper[4945]: I1008 15:52:48.944361 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qqhl\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-kube-api-access-8qqhl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-29p9d\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:49 crc kubenswrapper[4945]: I1008 15:52:49.039412 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:52:49 crc kubenswrapper[4945]: I1008 15:52:49.556087 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d"] Oct 08 15:52:49 crc kubenswrapper[4945]: I1008 15:52:49.560482 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:52:49 crc kubenswrapper[4945]: I1008 15:52:49.633503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" event={"ID":"b8610fc4-8948-4247-91b3-78ca4e5e1767","Type":"ContainerStarted","Data":"89a4bf8dc38f621cbff9dbfe960abfdbf26e29f2bbfb3d8309be276962a609c6"} Oct 08 15:52:50 crc kubenswrapper[4945]: I1008 15:52:50.024221 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:52:50 crc kubenswrapper[4945]: E1008 15:52:50.024700 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:52:50 crc kubenswrapper[4945]: I1008 15:52:50.658749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" event={"ID":"b8610fc4-8948-4247-91b3-78ca4e5e1767","Type":"ContainerStarted","Data":"086d252fdd4243f69f4c5fee65e7c03c500557177607e86950ade247370f9c54"} Oct 08 15:52:50 crc kubenswrapper[4945]: I1008 15:52:50.678050 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" podStartSLOduration=2.078979582 podStartE2EDuration="2.678033607s" podCreationTimestamp="2025-10-08 15:52:48 +0000 UTC" firstStartedPulling="2025-10-08 15:52:49.560143392 +0000 UTC m=+2378.914058293" lastFinishedPulling="2025-10-08 15:52:50.159197417 +0000 UTC m=+2379.513112318" observedRunningTime="2025-10-08 15:52:50.676680793 +0000 UTC m=+2380.030595724" watchObservedRunningTime="2025-10-08 15:52:50.678033607 +0000 UTC m=+2380.031948508" Oct 08 15:53:02 crc kubenswrapper[4945]: I1008 15:53:02.032487 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:53:02 crc kubenswrapper[4945]: E1008 15:53:02.034074 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:53:14 crc kubenswrapper[4945]: I1008 15:53:14.024171 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:53:14 crc kubenswrapper[4945]: E1008 15:53:14.026012 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:53:29 crc kubenswrapper[4945]: I1008 15:53:29.024623 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:53:29 crc kubenswrapper[4945]: E1008 15:53:29.025806 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:53:32 crc kubenswrapper[4945]: I1008 15:53:32.058082 4945 generic.go:334] "Generic (PLEG): container finished" podID="b8610fc4-8948-4247-91b3-78ca4e5e1767" containerID="086d252fdd4243f69f4c5fee65e7c03c500557177607e86950ade247370f9c54" exitCode=0 Oct 08 15:53:32 crc kubenswrapper[4945]: I1008 15:53:32.058156 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" event={"ID":"b8610fc4-8948-4247-91b3-78ca4e5e1767","Type":"ContainerDied","Data":"086d252fdd4243f69f4c5fee65e7c03c500557177607e86950ade247370f9c54"} Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.536579 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.664642 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ssh-key\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.664722 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-repo-setup-combined-ca-bundle\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.664824 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.664905 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665175 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ovn-combined-ca-bundle\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665224 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-inventory\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665285 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qqhl\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-kube-api-access-8qqhl\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665317 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665376 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-nova-combined-ca-bundle\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665471 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-bootstrap-combined-ca-bundle\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665552 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-telemetry-combined-ca-bundle\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-libvirt-combined-ca-bundle\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.665665 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-neutron-metadata-combined-ca-bundle\") pod \"b8610fc4-8948-4247-91b3-78ca4e5e1767\" (UID: \"b8610fc4-8948-4247-91b3-78ca4e5e1767\") " Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.673416 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.673593 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.673588 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.673699 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.676864 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.676905 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.677043 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.678706 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.679292 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.679985 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.685560 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.695172 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-kube-api-access-8qqhl" (OuterVolumeSpecName: "kube-api-access-8qqhl") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "kube-api-access-8qqhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.707348 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-inventory" (OuterVolumeSpecName: "inventory") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.736035 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8610fc4-8948-4247-91b3-78ca4e5e1767" (UID: "b8610fc4-8948-4247-91b3-78ca4e5e1767"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768319 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768349 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768361 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768371 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qqhl\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-kube-api-access-8qqhl\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768381 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768390 4945 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768400 4945 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768408 4945 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768416 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768426 4945 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768435 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768443 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768452 4945 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8610fc4-8948-4247-91b3-78ca4e5e1767-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:33 crc kubenswrapper[4945]: I1008 15:53:33.768462 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8610fc4-8948-4247-91b3-78ca4e5e1767-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.083039 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" event={"ID":"b8610fc4-8948-4247-91b3-78ca4e5e1767","Type":"ContainerDied","Data":"89a4bf8dc38f621cbff9dbfe960abfdbf26e29f2bbfb3d8309be276962a609c6"} Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.083148 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-29p9d" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.083104 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89a4bf8dc38f621cbff9dbfe960abfdbf26e29f2bbfb3d8309be276962a609c6" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.277924 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct"] Oct 08 15:53:34 crc kubenswrapper[4945]: E1008 15:53:34.278425 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8610fc4-8948-4247-91b3-78ca4e5e1767" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.278448 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8610fc4-8948-4247-91b3-78ca4e5e1767" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.278751 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8610fc4-8948-4247-91b3-78ca4e5e1767" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.279716 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.284870 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.284949 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.285005 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.285005 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.288708 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.300851 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct"] Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.379839 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.379907 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.380164 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.380282 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.380430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4psh\" (UniqueName: \"kubernetes.io/projected/d32f281f-0e3f-4ac5-8463-7bb261bc557e-kube-api-access-x4psh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.482707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.482765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.482812 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.482865 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.483627 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4psh\" (UniqueName: \"kubernetes.io/projected/d32f281f-0e3f-4ac5-8463-7bb261bc557e-kube-api-access-x4psh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.484022 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.488011 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.498062 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.501771 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.504383 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4psh\" (UniqueName: \"kubernetes.io/projected/d32f281f-0e3f-4ac5-8463-7bb261bc557e-kube-api-access-x4psh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wv9ct\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:34 crc kubenswrapper[4945]: I1008 15:53:34.602555 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:53:35 crc kubenswrapper[4945]: I1008 15:53:35.141193 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct"] Oct 08 15:53:36 crc kubenswrapper[4945]: I1008 15:53:36.105200 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" event={"ID":"d32f281f-0e3f-4ac5-8463-7bb261bc557e","Type":"ContainerStarted","Data":"fc3ae4cfd8124e94134f72d45c028d18095fd351dc7f9353e96055512feb159d"} Oct 08 15:53:36 crc kubenswrapper[4945]: I1008 15:53:36.105533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" event={"ID":"d32f281f-0e3f-4ac5-8463-7bb261bc557e","Type":"ContainerStarted","Data":"f2120e526faa812fdc9f242b30e2aa25dc02c84da244ece4c7527d8f7db8a1c4"} Oct 08 15:53:36 crc kubenswrapper[4945]: I1008 15:53:36.132821 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" podStartSLOduration=1.656601078 podStartE2EDuration="2.132793964s" podCreationTimestamp="2025-10-08 15:53:34 +0000 UTC" firstStartedPulling="2025-10-08 15:53:35.140850457 +0000 UTC m=+2424.494765358" lastFinishedPulling="2025-10-08 15:53:35.617043353 +0000 UTC m=+2424.970958244" observedRunningTime="2025-10-08 15:53:36.126081846 +0000 UTC m=+2425.479996757" watchObservedRunningTime="2025-10-08 15:53:36.132793964 +0000 UTC m=+2425.486708865" Oct 08 15:53:42 crc kubenswrapper[4945]: I1008 15:53:42.038571 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:53:42 crc kubenswrapper[4945]: E1008 15:53:42.039709 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:53:56 crc kubenswrapper[4945]: I1008 15:53:56.024446 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:53:56 crc kubenswrapper[4945]: E1008 15:53:56.025424 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:54:08 crc kubenswrapper[4945]: I1008 15:54:08.024100 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:54:08 crc kubenswrapper[4945]: E1008 15:54:08.024817 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:54:22 crc kubenswrapper[4945]: I1008 15:54:22.034470 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:54:22 crc kubenswrapper[4945]: E1008 15:54:22.035473 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:54:34 crc kubenswrapper[4945]: I1008 15:54:34.025431 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:54:34 crc kubenswrapper[4945]: E1008 15:54:34.026413 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:54:44 crc kubenswrapper[4945]: I1008 15:54:44.772486 4945 generic.go:334] "Generic (PLEG): container finished" podID="d32f281f-0e3f-4ac5-8463-7bb261bc557e" containerID="fc3ae4cfd8124e94134f72d45c028d18095fd351dc7f9353e96055512feb159d" exitCode=0 Oct 08 15:54:44 crc kubenswrapper[4945]: I1008 15:54:44.772549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" event={"ID":"d32f281f-0e3f-4ac5-8463-7bb261bc557e","Type":"ContainerDied","Data":"fc3ae4cfd8124e94134f72d45c028d18095fd351dc7f9353e96055512feb159d"} Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.234059 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.330072 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ssh-key\") pod \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.330968 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4psh\" (UniqueName: \"kubernetes.io/projected/d32f281f-0e3f-4ac5-8463-7bb261bc557e-kube-api-access-x4psh\") pod \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.331374 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-inventory\") pod \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.331414 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovn-combined-ca-bundle\") pod \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.331553 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovncontroller-config-0\") pod \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\" (UID: \"d32f281f-0e3f-4ac5-8463-7bb261bc557e\") " Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.337336 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d32f281f-0e3f-4ac5-8463-7bb261bc557e" (UID: "d32f281f-0e3f-4ac5-8463-7bb261bc557e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.337483 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32f281f-0e3f-4ac5-8463-7bb261bc557e-kube-api-access-x4psh" (OuterVolumeSpecName: "kube-api-access-x4psh") pod "d32f281f-0e3f-4ac5-8463-7bb261bc557e" (UID: "d32f281f-0e3f-4ac5-8463-7bb261bc557e"). InnerVolumeSpecName "kube-api-access-x4psh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.355904 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d32f281f-0e3f-4ac5-8463-7bb261bc557e" (UID: "d32f281f-0e3f-4ac5-8463-7bb261bc557e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.365823 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-inventory" (OuterVolumeSpecName: "inventory") pod "d32f281f-0e3f-4ac5-8463-7bb261bc557e" (UID: "d32f281f-0e3f-4ac5-8463-7bb261bc557e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.367852 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d32f281f-0e3f-4ac5-8463-7bb261bc557e" (UID: "d32f281f-0e3f-4ac5-8463-7bb261bc557e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.435164 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4psh\" (UniqueName: \"kubernetes.io/projected/d32f281f-0e3f-4ac5-8463-7bb261bc557e-kube-api-access-x4psh\") on node \"crc\" DevicePath \"\"" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.435420 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.435481 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.435536 4945 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.435601 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d32f281f-0e3f-4ac5-8463-7bb261bc557e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.794300 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" event={"ID":"d32f281f-0e3f-4ac5-8463-7bb261bc557e","Type":"ContainerDied","Data":"f2120e526faa812fdc9f242b30e2aa25dc02c84da244ece4c7527d8f7db8a1c4"} Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.794382 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wv9ct" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.794382 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2120e526faa812fdc9f242b30e2aa25dc02c84da244ece4c7527d8f7db8a1c4" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.904586 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p"] Oct 08 15:54:46 crc kubenswrapper[4945]: E1008 15:54:46.905026 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32f281f-0e3f-4ac5-8463-7bb261bc557e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.905043 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32f281f-0e3f-4ac5-8463-7bb261bc557e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.906147 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32f281f-0e3f-4ac5-8463-7bb261bc557e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.906949 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.909763 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.910315 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.910513 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.910599 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.910658 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.910787 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.922604 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p"] Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.957437 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.957525 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.957583 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwb22\" (UniqueName: \"kubernetes.io/projected/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-kube-api-access-qwb22\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.957618 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.957691 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:46 crc kubenswrapper[4945]: I1008 15:54:46.957742 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.024416 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:54:47 crc kubenswrapper[4945]: E1008 15:54:47.024746 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.059787 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.059897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.059962 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.060026 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwb22\" (UniqueName: \"kubernetes.io/projected/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-kube-api-access-qwb22\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.060053 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.060154 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.064360 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.064410 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.064445 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.065327 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.067011 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.078699 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwb22\" (UniqueName: \"kubernetes.io/projected/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-kube-api-access-qwb22\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.234299 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:54:47 crc kubenswrapper[4945]: I1008 15:54:47.869194 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p"] Oct 08 15:54:48 crc kubenswrapper[4945]: I1008 15:54:48.823500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" event={"ID":"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352","Type":"ContainerStarted","Data":"fa4ebcb033a52fad0f17e7625f0bd1ca661e48d97bf6970a7075d4f772afe391"} Oct 08 15:54:48 crc kubenswrapper[4945]: I1008 15:54:48.823840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" event={"ID":"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352","Type":"ContainerStarted","Data":"6c70926098f094aefb60ad5e724e1fecf4e6d25d3e7fdd4e27901d80ad464aa9"} Oct 08 15:54:48 crc kubenswrapper[4945]: I1008 15:54:48.847813 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" podStartSLOduration=2.306293387 podStartE2EDuration="2.847793746s" podCreationTimestamp="2025-10-08 15:54:46 +0000 UTC" firstStartedPulling="2025-10-08 15:54:47.878016647 +0000 UTC m=+2497.231931548" lastFinishedPulling="2025-10-08 15:54:48.419517006 +0000 UTC m=+2497.773431907" observedRunningTime="2025-10-08 15:54:48.843671943 +0000 UTC m=+2498.197586844" watchObservedRunningTime="2025-10-08 15:54:48.847793746 +0000 UTC m=+2498.201708647" Oct 08 15:54:58 crc kubenswrapper[4945]: I1008 15:54:58.026013 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:54:58 crc kubenswrapper[4945]: E1008 15:54:58.026930 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:55:10 crc kubenswrapper[4945]: I1008 15:55:10.025067 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:55:10 crc kubenswrapper[4945]: E1008 15:55:10.026065 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:55:21 crc kubenswrapper[4945]: I1008 15:55:21.024657 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:55:21 crc kubenswrapper[4945]: E1008 15:55:21.025420 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:55:33 crc kubenswrapper[4945]: I1008 15:55:33.024775 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:55:33 crc kubenswrapper[4945]: E1008 15:55:33.025814 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:55:39 crc kubenswrapper[4945]: I1008 15:55:39.316343 4945 generic.go:334] "Generic (PLEG): container finished" podID="0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" containerID="fa4ebcb033a52fad0f17e7625f0bd1ca661e48d97bf6970a7075d4f772afe391" exitCode=0 Oct 08 15:55:39 crc kubenswrapper[4945]: I1008 15:55:39.316467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" event={"ID":"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352","Type":"ContainerDied","Data":"fa4ebcb033a52fad0f17e7625f0bd1ca661e48d97bf6970a7075d4f772afe391"} Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.761874 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.847661 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-nova-metadata-neutron-config-0\") pod \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.847769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-ssh-key\") pod \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.847807 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-metadata-combined-ca-bundle\") pod \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.847903 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-ovn-metadata-agent-neutron-config-0\") pod \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.847955 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwb22\" (UniqueName: \"kubernetes.io/projected/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-kube-api-access-qwb22\") pod \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.848068 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-inventory\") pod \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\" (UID: \"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352\") " Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.855242 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-kube-api-access-qwb22" (OuterVolumeSpecName: "kube-api-access-qwb22") pod "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" (UID: "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352"). InnerVolumeSpecName "kube-api-access-qwb22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.880380 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" (UID: "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.892671 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" (UID: "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.893692 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" (UID: "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.900426 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-inventory" (OuterVolumeSpecName: "inventory") pod "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" (UID: "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.901972 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" (UID: "0c384ac4-a16d-4d25-b5e0-dc7eb33f7352"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.955294 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwb22\" (UniqueName: \"kubernetes.io/projected/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-kube-api-access-qwb22\") on node \"crc\" DevicePath \"\"" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.955684 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.955701 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.955717 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.955732 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:55:40 crc kubenswrapper[4945]: I1008 15:55:40.955748 4945 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0c384ac4-a16d-4d25-b5e0-dc7eb33f7352-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.346504 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" event={"ID":"0c384ac4-a16d-4d25-b5e0-dc7eb33f7352","Type":"ContainerDied","Data":"6c70926098f094aefb60ad5e724e1fecf4e6d25d3e7fdd4e27901d80ad464aa9"} Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.346544 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c70926098f094aefb60ad5e724e1fecf4e6d25d3e7fdd4e27901d80ad464aa9" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.346588 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.451870 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw"] Oct 08 15:55:41 crc kubenswrapper[4945]: E1008 15:55:41.452410 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.452433 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.452679 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c384ac4-a16d-4d25-b5e0-dc7eb33f7352" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.453566 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.461430 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.462240 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.462253 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.462325 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.462683 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.464714 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw"] Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.568182 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7nmk\" (UniqueName: \"kubernetes.io/projected/8a77f15d-769f-493c-b410-eb13c9df8fe6-kube-api-access-x7nmk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.568242 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.568441 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.568546 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.568619 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.670903 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.670998 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.671054 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.671125 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7nmk\" (UniqueName: \"kubernetes.io/projected/8a77f15d-769f-493c-b410-eb13c9df8fe6-kube-api-access-x7nmk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.671148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.675718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.676136 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.677435 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.681514 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.689130 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7nmk\" (UniqueName: \"kubernetes.io/projected/8a77f15d-769f-493c-b410-eb13c9df8fe6-kube-api-access-x7nmk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-76mxw\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:41 crc kubenswrapper[4945]: I1008 15:55:41.775054 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 15:55:42 crc kubenswrapper[4945]: I1008 15:55:42.363374 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw"] Oct 08 15:55:43 crc kubenswrapper[4945]: I1008 15:55:43.374554 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" event={"ID":"8a77f15d-769f-493c-b410-eb13c9df8fe6","Type":"ContainerStarted","Data":"ecc9550dce488541b4e634982b001dcf7c340dc7d092c13f832f59cc712821b7"} Oct 08 15:55:43 crc kubenswrapper[4945]: I1008 15:55:43.375060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" event={"ID":"8a77f15d-769f-493c-b410-eb13c9df8fe6","Type":"ContainerStarted","Data":"101a42644c7632934af4de837ef78af289ccd9c706683eee44cbbb241382f359"} Oct 08 15:55:43 crc kubenswrapper[4945]: I1008 15:55:43.398076 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" podStartSLOduration=1.911194052 podStartE2EDuration="2.398054327s" podCreationTimestamp="2025-10-08 15:55:41 +0000 UTC" firstStartedPulling="2025-10-08 15:55:42.371875528 +0000 UTC m=+2551.725790429" lastFinishedPulling="2025-10-08 15:55:42.858735793 +0000 UTC m=+2552.212650704" observedRunningTime="2025-10-08 15:55:43.390779144 +0000 UTC m=+2552.744694045" watchObservedRunningTime="2025-10-08 15:55:43.398054327 +0000 UTC m=+2552.751969228" Oct 08 15:55:48 crc kubenswrapper[4945]: I1008 15:55:48.024119 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:55:48 crc kubenswrapper[4945]: E1008 15:55:48.024825 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:56:01 crc kubenswrapper[4945]: I1008 15:56:01.024080 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:56:01 crc kubenswrapper[4945]: E1008 15:56:01.024828 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:56:15 crc kubenswrapper[4945]: I1008 15:56:15.024596 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:56:15 crc kubenswrapper[4945]: E1008 15:56:15.025570 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 15:56:30 crc kubenswrapper[4945]: I1008 15:56:30.024733 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 15:56:30 crc kubenswrapper[4945]: I1008 15:56:30.877645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"cbd03bbb5effbe7203ee78f496768b2e3ade6cfa776d6ff5e8810e49735e8777"} Oct 08 15:58:49 crc kubenswrapper[4945]: I1008 15:58:49.184591 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:58:49 crc kubenswrapper[4945]: I1008 15:58:49.185221 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:59:19 crc kubenswrapper[4945]: I1008 15:59:19.183857 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:59:19 crc kubenswrapper[4945]: I1008 15:59:19.184747 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.073362 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pfv54"] Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.075912 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.094105 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pfv54"] Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.263636 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-utilities\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.263707 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-catalog-content\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.263822 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxjdj\" (UniqueName: \"kubernetes.io/projected/17060f70-419b-4186-a177-05cb4648861c-kube-api-access-zxjdj\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.366186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-utilities\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.366239 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-catalog-content\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.366332 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxjdj\" (UniqueName: \"kubernetes.io/projected/17060f70-419b-4186-a177-05cb4648861c-kube-api-access-zxjdj\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.366720 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-utilities\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.366732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-catalog-content\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.384930 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxjdj\" (UniqueName: \"kubernetes.io/projected/17060f70-419b-4186-a177-05cb4648861c-kube-api-access-zxjdj\") pod \"community-operators-pfv54\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.406770 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:23 crc kubenswrapper[4945]: I1008 15:59:23.931732 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pfv54"] Oct 08 15:59:24 crc kubenswrapper[4945]: I1008 15:59:24.605532 4945 generic.go:334] "Generic (PLEG): container finished" podID="17060f70-419b-4186-a177-05cb4648861c" containerID="a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e" exitCode=0 Oct 08 15:59:24 crc kubenswrapper[4945]: I1008 15:59:24.605816 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pfv54" event={"ID":"17060f70-419b-4186-a177-05cb4648861c","Type":"ContainerDied","Data":"a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e"} Oct 08 15:59:24 crc kubenswrapper[4945]: I1008 15:59:24.605847 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pfv54" event={"ID":"17060f70-419b-4186-a177-05cb4648861c","Type":"ContainerStarted","Data":"6117613663d99fc18558750707d0f22a9cd915f5edd658bdb79c373e570fe0b7"} Oct 08 15:59:24 crc kubenswrapper[4945]: I1008 15:59:24.610156 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:59:26 crc kubenswrapper[4945]: I1008 15:59:26.626090 4945 generic.go:334] "Generic (PLEG): container finished" podID="17060f70-419b-4186-a177-05cb4648861c" containerID="78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046" exitCode=0 Oct 08 15:59:26 crc kubenswrapper[4945]: I1008 15:59:26.626206 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pfv54" event={"ID":"17060f70-419b-4186-a177-05cb4648861c","Type":"ContainerDied","Data":"78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046"} Oct 08 15:59:27 crc kubenswrapper[4945]: I1008 15:59:27.636358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pfv54" event={"ID":"17060f70-419b-4186-a177-05cb4648861c","Type":"ContainerStarted","Data":"1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de"} Oct 08 15:59:27 crc kubenswrapper[4945]: I1008 15:59:27.657218 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pfv54" podStartSLOduration=1.977060289 podStartE2EDuration="4.657195965s" podCreationTimestamp="2025-10-08 15:59:23 +0000 UTC" firstStartedPulling="2025-10-08 15:59:24.609877627 +0000 UTC m=+2773.963792528" lastFinishedPulling="2025-10-08 15:59:27.290013313 +0000 UTC m=+2776.643928204" observedRunningTime="2025-10-08 15:59:27.653148023 +0000 UTC m=+2777.007062924" watchObservedRunningTime="2025-10-08 15:59:27.657195965 +0000 UTC m=+2777.011110876" Oct 08 15:59:33 crc kubenswrapper[4945]: I1008 15:59:33.407492 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:33 crc kubenswrapper[4945]: I1008 15:59:33.408085 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:33 crc kubenswrapper[4945]: I1008 15:59:33.453570 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:33 crc kubenswrapper[4945]: I1008 15:59:33.731875 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:33 crc kubenswrapper[4945]: I1008 15:59:33.777001 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pfv54"] Oct 08 15:59:35 crc kubenswrapper[4945]: I1008 15:59:35.709478 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pfv54" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="registry-server" containerID="cri-o://1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de" gracePeriod=2 Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.159915 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.228505 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxjdj\" (UniqueName: \"kubernetes.io/projected/17060f70-419b-4186-a177-05cb4648861c-kube-api-access-zxjdj\") pod \"17060f70-419b-4186-a177-05cb4648861c\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.228556 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-utilities\") pod \"17060f70-419b-4186-a177-05cb4648861c\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.228762 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-catalog-content\") pod \"17060f70-419b-4186-a177-05cb4648861c\" (UID: \"17060f70-419b-4186-a177-05cb4648861c\") " Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.229431 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-utilities" (OuterVolumeSpecName: "utilities") pod "17060f70-419b-4186-a177-05cb4648861c" (UID: "17060f70-419b-4186-a177-05cb4648861c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.234596 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17060f70-419b-4186-a177-05cb4648861c-kube-api-access-zxjdj" (OuterVolumeSpecName: "kube-api-access-zxjdj") pod "17060f70-419b-4186-a177-05cb4648861c" (UID: "17060f70-419b-4186-a177-05cb4648861c"). InnerVolumeSpecName "kube-api-access-zxjdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.277702 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17060f70-419b-4186-a177-05cb4648861c" (UID: "17060f70-419b-4186-a177-05cb4648861c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.330984 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxjdj\" (UniqueName: \"kubernetes.io/projected/17060f70-419b-4186-a177-05cb4648861c-kube-api-access-zxjdj\") on node \"crc\" DevicePath \"\"" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.331029 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.331041 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17060f70-419b-4186-a177-05cb4648861c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.722368 4945 generic.go:334] "Generic (PLEG): container finished" podID="17060f70-419b-4186-a177-05cb4648861c" containerID="1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de" exitCode=0 Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.722403 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pfv54" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.722422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pfv54" event={"ID":"17060f70-419b-4186-a177-05cb4648861c","Type":"ContainerDied","Data":"1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de"} Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.722959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pfv54" event={"ID":"17060f70-419b-4186-a177-05cb4648861c","Type":"ContainerDied","Data":"6117613663d99fc18558750707d0f22a9cd915f5edd658bdb79c373e570fe0b7"} Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.722986 4945 scope.go:117] "RemoveContainer" containerID="1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.747558 4945 scope.go:117] "RemoveContainer" containerID="78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.767242 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pfv54"] Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.778301 4945 scope.go:117] "RemoveContainer" containerID="a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.778467 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pfv54"] Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.830567 4945 scope.go:117] "RemoveContainer" containerID="1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de" Oct 08 15:59:36 crc kubenswrapper[4945]: E1008 15:59:36.831074 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de\": container with ID starting with 1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de not found: ID does not exist" containerID="1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.831145 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de"} err="failed to get container status \"1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de\": rpc error: code = NotFound desc = could not find container \"1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de\": container with ID starting with 1b325296b411932ac3815a1c53704fc92f49ae0982ca2f465dc35e256c4161de not found: ID does not exist" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.831176 4945 scope.go:117] "RemoveContainer" containerID="78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046" Oct 08 15:59:36 crc kubenswrapper[4945]: E1008 15:59:36.831690 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046\": container with ID starting with 78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046 not found: ID does not exist" containerID="78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.831742 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046"} err="failed to get container status \"78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046\": rpc error: code = NotFound desc = could not find container \"78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046\": container with ID starting with 78faea6d93051a258dfe84c420e707ad740401bebab5495a21a7904bc5595046 not found: ID does not exist" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.831778 4945 scope.go:117] "RemoveContainer" containerID="a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e" Oct 08 15:59:36 crc kubenswrapper[4945]: E1008 15:59:36.832214 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e\": container with ID starting with a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e not found: ID does not exist" containerID="a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e" Oct 08 15:59:36 crc kubenswrapper[4945]: I1008 15:59:36.832248 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e"} err="failed to get container status \"a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e\": rpc error: code = NotFound desc = could not find container \"a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e\": container with ID starting with a848b4661dd2e94e549e3ba36e2e1ee779c8ea5d3daacfc0357683b04b69b09e not found: ID does not exist" Oct 08 15:59:38 crc kubenswrapper[4945]: I1008 15:59:38.037414 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17060f70-419b-4186-a177-05cb4648861c" path="/var/lib/kubelet/pods/17060f70-419b-4186-a177-05cb4648861c/volumes" Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.183826 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.184455 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.184506 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.185280 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cbd03bbb5effbe7203ee78f496768b2e3ade6cfa776d6ff5e8810e49735e8777"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.185348 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://cbd03bbb5effbe7203ee78f496768b2e3ade6cfa776d6ff5e8810e49735e8777" gracePeriod=600 Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.863928 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="cbd03bbb5effbe7203ee78f496768b2e3ade6cfa776d6ff5e8810e49735e8777" exitCode=0 Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.863969 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"cbd03bbb5effbe7203ee78f496768b2e3ade6cfa776d6ff5e8810e49735e8777"} Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.864361 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609"} Oct 08 15:59:49 crc kubenswrapper[4945]: I1008 15:59:49.864394 4945 scope.go:117] "RemoveContainer" containerID="6749880d769cea3b8f58b678d4ff14d7d470428ffbd90447fd7ee93be12b8bb4" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.155214 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx"] Oct 08 16:00:00 crc kubenswrapper[4945]: E1008 16:00:00.156674 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="extract-content" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.156702 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="extract-content" Oct 08 16:00:00 crc kubenswrapper[4945]: E1008 16:00:00.156761 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="registry-server" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.156777 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="registry-server" Oct 08 16:00:00 crc kubenswrapper[4945]: E1008 16:00:00.156824 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="extract-utilities" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.156838 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="extract-utilities" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.157223 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="17060f70-419b-4186-a177-05cb4648861c" containerName="registry-server" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.158501 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.160883 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.161250 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.167861 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx"] Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.220751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scqvg\" (UniqueName: \"kubernetes.io/projected/a696c238-b2c9-4d8b-9977-08ea3c148dbe-kube-api-access-scqvg\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.221322 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a696c238-b2c9-4d8b-9977-08ea3c148dbe-config-volume\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.221512 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a696c238-b2c9-4d8b-9977-08ea3c148dbe-secret-volume\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.324009 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a696c238-b2c9-4d8b-9977-08ea3c148dbe-secret-volume\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.324268 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scqvg\" (UniqueName: \"kubernetes.io/projected/a696c238-b2c9-4d8b-9977-08ea3c148dbe-kube-api-access-scqvg\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.324327 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a696c238-b2c9-4d8b-9977-08ea3c148dbe-config-volume\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.325410 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a696c238-b2c9-4d8b-9977-08ea3c148dbe-config-volume\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.330593 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a696c238-b2c9-4d8b-9977-08ea3c148dbe-secret-volume\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.341753 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scqvg\" (UniqueName: \"kubernetes.io/projected/a696c238-b2c9-4d8b-9977-08ea3c148dbe-kube-api-access-scqvg\") pod \"collect-profiles-29332320-rwnqx\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.483642 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.935851 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx"] Oct 08 16:00:00 crc kubenswrapper[4945]: W1008 16:00:00.938617 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda696c238_b2c9_4d8b_9977_08ea3c148dbe.slice/crio-9103c8274dc614d46baf4544b8377bac1b742b9deee29255df604054a68e52ea WatchSource:0}: Error finding container 9103c8274dc614d46baf4544b8377bac1b742b9deee29255df604054a68e52ea: Status 404 returned error can't find the container with id 9103c8274dc614d46baf4544b8377bac1b742b9deee29255df604054a68e52ea Oct 08 16:00:00 crc kubenswrapper[4945]: I1008 16:00:00.989204 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" event={"ID":"a696c238-b2c9-4d8b-9977-08ea3c148dbe","Type":"ContainerStarted","Data":"9103c8274dc614d46baf4544b8377bac1b742b9deee29255df604054a68e52ea"} Oct 08 16:00:01 crc kubenswrapper[4945]: I1008 16:00:01.999526 4945 generic.go:334] "Generic (PLEG): container finished" podID="a696c238-b2c9-4d8b-9977-08ea3c148dbe" containerID="e68e9c13cde3d93ea02f97d78c0e8185e826e68d0087f62081435d2994c4ec19" exitCode=0 Oct 08 16:00:01 crc kubenswrapper[4945]: I1008 16:00:01.999629 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" event={"ID":"a696c238-b2c9-4d8b-9977-08ea3c148dbe","Type":"ContainerDied","Data":"e68e9c13cde3d93ea02f97d78c0e8185e826e68d0087f62081435d2994c4ec19"} Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.669484 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8k56d"] Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.672466 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.694033 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8k56d"] Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.779055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-utilities\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.779426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbs5p\" (UniqueName: \"kubernetes.io/projected/1ae60c2a-f368-4a15-836d-38b76ddb59ba-kube-api-access-dbs5p\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.779545 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-catalog-content\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.881544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-utilities\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.881615 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbs5p\" (UniqueName: \"kubernetes.io/projected/1ae60c2a-f368-4a15-836d-38b76ddb59ba-kube-api-access-dbs5p\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.881651 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-catalog-content\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.882067 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-utilities\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.882179 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-catalog-content\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:02 crc kubenswrapper[4945]: I1008 16:00:02.904124 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbs5p\" (UniqueName: \"kubernetes.io/projected/1ae60c2a-f368-4a15-836d-38b76ddb59ba-kube-api-access-dbs5p\") pod \"certified-operators-8k56d\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.007818 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.413682 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.493692 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scqvg\" (UniqueName: \"kubernetes.io/projected/a696c238-b2c9-4d8b-9977-08ea3c148dbe-kube-api-access-scqvg\") pod \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.493818 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a696c238-b2c9-4d8b-9977-08ea3c148dbe-secret-volume\") pod \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.494021 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a696c238-b2c9-4d8b-9977-08ea3c148dbe-config-volume\") pod \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\" (UID: \"a696c238-b2c9-4d8b-9977-08ea3c148dbe\") " Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.494568 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a696c238-b2c9-4d8b-9977-08ea3c148dbe-config-volume" (OuterVolumeSpecName: "config-volume") pod "a696c238-b2c9-4d8b-9977-08ea3c148dbe" (UID: "a696c238-b2c9-4d8b-9977-08ea3c148dbe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.499012 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a696c238-b2c9-4d8b-9977-08ea3c148dbe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a696c238-b2c9-4d8b-9977-08ea3c148dbe" (UID: "a696c238-b2c9-4d8b-9977-08ea3c148dbe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.499519 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a696c238-b2c9-4d8b-9977-08ea3c148dbe-kube-api-access-scqvg" (OuterVolumeSpecName: "kube-api-access-scqvg") pod "a696c238-b2c9-4d8b-9977-08ea3c148dbe" (UID: "a696c238-b2c9-4d8b-9977-08ea3c148dbe"). InnerVolumeSpecName "kube-api-access-scqvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.595729 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a696c238-b2c9-4d8b-9977-08ea3c148dbe-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.595756 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a696c238-b2c9-4d8b-9977-08ea3c148dbe-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.595769 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scqvg\" (UniqueName: \"kubernetes.io/projected/a696c238-b2c9-4d8b-9977-08ea3c148dbe-kube-api-access-scqvg\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:03 crc kubenswrapper[4945]: I1008 16:00:03.611210 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8k56d"] Oct 08 16:00:03 crc kubenswrapper[4945]: W1008 16:00:03.611526 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ae60c2a_f368_4a15_836d_38b76ddb59ba.slice/crio-ebf3c70b303ec32b42a9ae2a40354ef2ccdcd1cc95a65353f9b7afee06af4073 WatchSource:0}: Error finding container ebf3c70b303ec32b42a9ae2a40354ef2ccdcd1cc95a65353f9b7afee06af4073: Status 404 returned error can't find the container with id ebf3c70b303ec32b42a9ae2a40354ef2ccdcd1cc95a65353f9b7afee06af4073 Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.017973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" event={"ID":"a696c238-b2c9-4d8b-9977-08ea3c148dbe","Type":"ContainerDied","Data":"9103c8274dc614d46baf4544b8377bac1b742b9deee29255df604054a68e52ea"} Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.018331 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9103c8274dc614d46baf4544b8377bac1b742b9deee29255df604054a68e52ea" Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.018011 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx" Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.022065 4945 generic.go:334] "Generic (PLEG): container finished" podID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerID="f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e" exitCode=0 Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.022144 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k56d" event={"ID":"1ae60c2a-f368-4a15-836d-38b76ddb59ba","Type":"ContainerDied","Data":"f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e"} Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.022189 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k56d" event={"ID":"1ae60c2a-f368-4a15-836d-38b76ddb59ba","Type":"ContainerStarted","Data":"ebf3c70b303ec32b42a9ae2a40354ef2ccdcd1cc95a65353f9b7afee06af4073"} Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.488718 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp"] Oct 08 16:00:04 crc kubenswrapper[4945]: I1008 16:00:04.497892 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-wplnp"] Oct 08 16:00:06 crc kubenswrapper[4945]: I1008 16:00:06.044084 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ba793aa-5158-46a3-8ea6-1e91389524d1" path="/var/lib/kubelet/pods/2ba793aa-5158-46a3-8ea6-1e91389524d1/volumes" Oct 08 16:00:06 crc kubenswrapper[4945]: I1008 16:00:06.048508 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k56d" event={"ID":"1ae60c2a-f368-4a15-836d-38b76ddb59ba","Type":"ContainerStarted","Data":"e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866"} Oct 08 16:00:08 crc kubenswrapper[4945]: I1008 16:00:08.063623 4945 generic.go:334] "Generic (PLEG): container finished" podID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerID="e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866" exitCode=0 Oct 08 16:00:08 crc kubenswrapper[4945]: I1008 16:00:08.063705 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k56d" event={"ID":"1ae60c2a-f368-4a15-836d-38b76ddb59ba","Type":"ContainerDied","Data":"e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866"} Oct 08 16:00:09 crc kubenswrapper[4945]: I1008 16:00:09.080421 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k56d" event={"ID":"1ae60c2a-f368-4a15-836d-38b76ddb59ba","Type":"ContainerStarted","Data":"1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77"} Oct 08 16:00:09 crc kubenswrapper[4945]: I1008 16:00:09.108315 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8k56d" podStartSLOduration=2.571290314 podStartE2EDuration="7.108293635s" podCreationTimestamp="2025-10-08 16:00:02 +0000 UTC" firstStartedPulling="2025-10-08 16:00:04.02425568 +0000 UTC m=+2813.378170601" lastFinishedPulling="2025-10-08 16:00:08.561259021 +0000 UTC m=+2817.915173922" observedRunningTime="2025-10-08 16:00:09.10251239 +0000 UTC m=+2818.456427331" watchObservedRunningTime="2025-10-08 16:00:09.108293635 +0000 UTC m=+2818.462208526" Oct 08 16:00:13 crc kubenswrapper[4945]: I1008 16:00:13.008282 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:13 crc kubenswrapper[4945]: I1008 16:00:13.009209 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:13 crc kubenswrapper[4945]: I1008 16:00:13.090562 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:13 crc kubenswrapper[4945]: I1008 16:00:13.125153 4945 generic.go:334] "Generic (PLEG): container finished" podID="8a77f15d-769f-493c-b410-eb13c9df8fe6" containerID="ecc9550dce488541b4e634982b001dcf7c340dc7d092c13f832f59cc712821b7" exitCode=0 Oct 08 16:00:13 crc kubenswrapper[4945]: I1008 16:00:13.125212 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" event={"ID":"8a77f15d-769f-493c-b410-eb13c9df8fe6","Type":"ContainerDied","Data":"ecc9550dce488541b4e634982b001dcf7c340dc7d092c13f832f59cc712821b7"} Oct 08 16:00:13 crc kubenswrapper[4945]: I1008 16:00:13.180388 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:13 crc kubenswrapper[4945]: I1008 16:00:13.335911 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8k56d"] Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.579145 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.616397 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-ssh-key\") pod \"8a77f15d-769f-493c-b410-eb13c9df8fe6\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.616478 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-combined-ca-bundle\") pod \"8a77f15d-769f-493c-b410-eb13c9df8fe6\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.616555 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7nmk\" (UniqueName: \"kubernetes.io/projected/8a77f15d-769f-493c-b410-eb13c9df8fe6-kube-api-access-x7nmk\") pod \"8a77f15d-769f-493c-b410-eb13c9df8fe6\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.616688 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-inventory\") pod \"8a77f15d-769f-493c-b410-eb13c9df8fe6\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.616720 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-secret-0\") pod \"8a77f15d-769f-493c-b410-eb13c9df8fe6\" (UID: \"8a77f15d-769f-493c-b410-eb13c9df8fe6\") " Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.625691 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8a77f15d-769f-493c-b410-eb13c9df8fe6" (UID: "8a77f15d-769f-493c-b410-eb13c9df8fe6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.625717 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a77f15d-769f-493c-b410-eb13c9df8fe6-kube-api-access-x7nmk" (OuterVolumeSpecName: "kube-api-access-x7nmk") pod "8a77f15d-769f-493c-b410-eb13c9df8fe6" (UID: "8a77f15d-769f-493c-b410-eb13c9df8fe6"). InnerVolumeSpecName "kube-api-access-x7nmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.645718 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-inventory" (OuterVolumeSpecName: "inventory") pod "8a77f15d-769f-493c-b410-eb13c9df8fe6" (UID: "8a77f15d-769f-493c-b410-eb13c9df8fe6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.647420 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "8a77f15d-769f-493c-b410-eb13c9df8fe6" (UID: "8a77f15d-769f-493c-b410-eb13c9df8fe6"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.659799 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a77f15d-769f-493c-b410-eb13c9df8fe6" (UID: "8a77f15d-769f-493c-b410-eb13c9df8fe6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.719201 4945 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.719233 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7nmk\" (UniqueName: \"kubernetes.io/projected/8a77f15d-769f-493c-b410-eb13c9df8fe6-kube-api-access-x7nmk\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.719242 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.719253 4945 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:14 crc kubenswrapper[4945]: I1008 16:00:14.719261 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a77f15d-769f-493c-b410-eb13c9df8fe6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.143978 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" event={"ID":"8a77f15d-769f-493c-b410-eb13c9df8fe6","Type":"ContainerDied","Data":"101a42644c7632934af4de837ef78af289ccd9c706683eee44cbbb241382f359"} Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.144030 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="101a42644c7632934af4de837ef78af289ccd9c706683eee44cbbb241382f359" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.144039 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-76mxw" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.144125 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8k56d" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="registry-server" containerID="cri-o://1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77" gracePeriod=2 Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.255057 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75"] Oct 08 16:00:15 crc kubenswrapper[4945]: E1008 16:00:15.255799 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a696c238-b2c9-4d8b-9977-08ea3c148dbe" containerName="collect-profiles" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.255818 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a696c238-b2c9-4d8b-9977-08ea3c148dbe" containerName="collect-profiles" Oct 08 16:00:15 crc kubenswrapper[4945]: E1008 16:00:15.255840 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a77f15d-769f-493c-b410-eb13c9df8fe6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.255847 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a77f15d-769f-493c-b410-eb13c9df8fe6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.256055 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a77f15d-769f-493c-b410-eb13c9df8fe6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.256072 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a696c238-b2c9-4d8b-9977-08ea3c148dbe" containerName="collect-profiles" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.256760 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.259084 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.259796 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.260100 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.261299 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.261506 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.261642 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.261795 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.264021 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75"] Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.330736 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.330861 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.330900 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.330950 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.330977 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.331007 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.331058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.331087 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr6jr\" (UniqueName: \"kubernetes.io/projected/8b121934-fc40-487d-91fe-cdc91272f2dd-kube-api-access-nr6jr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.331159 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433340 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433535 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433553 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433618 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433643 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433682 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.433704 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr6jr\" (UniqueName: \"kubernetes.io/projected/8b121934-fc40-487d-91fe-cdc91272f2dd-kube-api-access-nr6jr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.436541 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.437944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.438394 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.439019 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.439500 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.439744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.439942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.441927 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.453240 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr6jr\" (UniqueName: \"kubernetes.io/projected/8b121934-fc40-487d-91fe-cdc91272f2dd-kube-api-access-nr6jr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mpr75\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.570173 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.622246 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.636862 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbs5p\" (UniqueName: \"kubernetes.io/projected/1ae60c2a-f368-4a15-836d-38b76ddb59ba-kube-api-access-dbs5p\") pod \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.636909 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-catalog-content\") pod \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.636946 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-utilities\") pod \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\" (UID: \"1ae60c2a-f368-4a15-836d-38b76ddb59ba\") " Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.638263 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-utilities" (OuterVolumeSpecName: "utilities") pod "1ae60c2a-f368-4a15-836d-38b76ddb59ba" (UID: "1ae60c2a-f368-4a15-836d-38b76ddb59ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.641091 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ae60c2a-f368-4a15-836d-38b76ddb59ba-kube-api-access-dbs5p" (OuterVolumeSpecName: "kube-api-access-dbs5p") pod "1ae60c2a-f368-4a15-836d-38b76ddb59ba" (UID: "1ae60c2a-f368-4a15-836d-38b76ddb59ba"). InnerVolumeSpecName "kube-api-access-dbs5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.691523 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ae60c2a-f368-4a15-836d-38b76ddb59ba" (UID: "1ae60c2a-f368-4a15-836d-38b76ddb59ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.739470 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbs5p\" (UniqueName: \"kubernetes.io/projected/1ae60c2a-f368-4a15-836d-38b76ddb59ba-kube-api-access-dbs5p\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.739508 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:15 crc kubenswrapper[4945]: I1008 16:00:15.739517 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ae60c2a-f368-4a15-836d-38b76ddb59ba-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.144057 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75"] Oct 08 16:00:16 crc kubenswrapper[4945]: W1008 16:00:16.150300 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b121934_fc40_487d_91fe_cdc91272f2dd.slice/crio-96788fa4c49e2863b856ab352ac1a386b38cc6c844a8cbf941cfcac822c5f220 WatchSource:0}: Error finding container 96788fa4c49e2863b856ab352ac1a386b38cc6c844a8cbf941cfcac822c5f220: Status 404 returned error can't find the container with id 96788fa4c49e2863b856ab352ac1a386b38cc6c844a8cbf941cfcac822c5f220 Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.157077 4945 generic.go:334] "Generic (PLEG): container finished" podID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerID="1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77" exitCode=0 Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.157161 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k56d" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.157159 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k56d" event={"ID":"1ae60c2a-f368-4a15-836d-38b76ddb59ba","Type":"ContainerDied","Data":"1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77"} Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.157219 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k56d" event={"ID":"1ae60c2a-f368-4a15-836d-38b76ddb59ba","Type":"ContainerDied","Data":"ebf3c70b303ec32b42a9ae2a40354ef2ccdcd1cc95a65353f9b7afee06af4073"} Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.157243 4945 scope.go:117] "RemoveContainer" containerID="1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.214681 4945 scope.go:117] "RemoveContainer" containerID="e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.215574 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8k56d"] Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.225461 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8k56d"] Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.238003 4945 scope.go:117] "RemoveContainer" containerID="f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.255168 4945 scope.go:117] "RemoveContainer" containerID="1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77" Oct 08 16:00:16 crc kubenswrapper[4945]: E1008 16:00:16.255671 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77\": container with ID starting with 1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77 not found: ID does not exist" containerID="1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.255729 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77"} err="failed to get container status \"1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77\": rpc error: code = NotFound desc = could not find container \"1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77\": container with ID starting with 1505413decf608bf3e653722276ffb78e86addfcec2dbcd08cb514c3b0e58c77 not found: ID does not exist" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.255773 4945 scope.go:117] "RemoveContainer" containerID="e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866" Oct 08 16:00:16 crc kubenswrapper[4945]: E1008 16:00:16.256277 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866\": container with ID starting with e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866 not found: ID does not exist" containerID="e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.256314 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866"} err="failed to get container status \"e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866\": rpc error: code = NotFound desc = could not find container \"e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866\": container with ID starting with e6ad3e9da6220cd292c21ba9da9fe8f50c5c7f8d08f1e23d16526ea7a3739866 not found: ID does not exist" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.256338 4945 scope.go:117] "RemoveContainer" containerID="f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e" Oct 08 16:00:16 crc kubenswrapper[4945]: E1008 16:00:16.256631 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e\": container with ID starting with f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e not found: ID does not exist" containerID="f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e" Oct 08 16:00:16 crc kubenswrapper[4945]: I1008 16:00:16.256655 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e"} err="failed to get container status \"f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e\": rpc error: code = NotFound desc = could not find container \"f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e\": container with ID starting with f5e3de25309c121a274d4ba1aab058c69fea878036e2db38adaf071ba07eb58e not found: ID does not exist" Oct 08 16:00:17 crc kubenswrapper[4945]: I1008 16:00:17.171748 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" event={"ID":"8b121934-fc40-487d-91fe-cdc91272f2dd","Type":"ContainerStarted","Data":"96788fa4c49e2863b856ab352ac1a386b38cc6c844a8cbf941cfcac822c5f220"} Oct 08 16:00:18 crc kubenswrapper[4945]: I1008 16:00:18.040532 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" path="/var/lib/kubelet/pods/1ae60c2a-f368-4a15-836d-38b76ddb59ba/volumes" Oct 08 16:00:18 crc kubenswrapper[4945]: I1008 16:00:18.186034 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" event={"ID":"8b121934-fc40-487d-91fe-cdc91272f2dd","Type":"ContainerStarted","Data":"a1ca5691ca7108e29e13a477a810732c62df29f5690b442f7c1a1084d7267cf5"} Oct 08 16:00:18 crc kubenswrapper[4945]: I1008 16:00:18.206612 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" podStartSLOduration=2.312024037 podStartE2EDuration="3.206590729s" podCreationTimestamp="2025-10-08 16:00:15 +0000 UTC" firstStartedPulling="2025-10-08 16:00:16.152793221 +0000 UTC m=+2825.506708122" lastFinishedPulling="2025-10-08 16:00:17.047359873 +0000 UTC m=+2826.401274814" observedRunningTime="2025-10-08 16:00:18.205156103 +0000 UTC m=+2827.559071004" watchObservedRunningTime="2025-10-08 16:00:18.206590729 +0000 UTC m=+2827.560505630" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.881468 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rbtmn"] Oct 08 16:00:29 crc kubenswrapper[4945]: E1008 16:00:29.882473 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="registry-server" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.882487 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="registry-server" Oct 08 16:00:29 crc kubenswrapper[4945]: E1008 16:00:29.882506 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="extract-content" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.882512 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="extract-content" Oct 08 16:00:29 crc kubenswrapper[4945]: E1008 16:00:29.882519 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="extract-utilities" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.882525 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="extract-utilities" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.882712 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae60c2a-f368-4a15-836d-38b76ddb59ba" containerName="registry-server" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.884383 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.899776 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbtmn"] Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.946058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-catalog-content\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.946169 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-utilities\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:29 crc kubenswrapper[4945]: I1008 16:00:29.946213 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7zlg\" (UniqueName: \"kubernetes.io/projected/3f53d421-a1b5-4495-9b35-d5e0805076e1-kube-api-access-f7zlg\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.047890 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-catalog-content\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.047982 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-utilities\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.048030 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7zlg\" (UniqueName: \"kubernetes.io/projected/3f53d421-a1b5-4495-9b35-d5e0805076e1-kube-api-access-f7zlg\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.048491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-catalog-content\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.048833 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-utilities\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.066236 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7zlg\" (UniqueName: \"kubernetes.io/projected/3f53d421-a1b5-4495-9b35-d5e0805076e1-kube-api-access-f7zlg\") pod \"redhat-marketplace-rbtmn\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.204236 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:30 crc kubenswrapper[4945]: I1008 16:00:30.648303 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbtmn"] Oct 08 16:00:30 crc kubenswrapper[4945]: W1008 16:00:30.650582 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f53d421_a1b5_4495_9b35_d5e0805076e1.slice/crio-7a746ac016e96dd99c24350efadd16f407cf9543b3b3b2d8c95c6da6397888f3 WatchSource:0}: Error finding container 7a746ac016e96dd99c24350efadd16f407cf9543b3b3b2d8c95c6da6397888f3: Status 404 returned error can't find the container with id 7a746ac016e96dd99c24350efadd16f407cf9543b3b3b2d8c95c6da6397888f3 Oct 08 16:00:31 crc kubenswrapper[4945]: I1008 16:00:31.331536 4945 generic.go:334] "Generic (PLEG): container finished" podID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerID="57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682" exitCode=0 Oct 08 16:00:31 crc kubenswrapper[4945]: I1008 16:00:31.331678 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbtmn" event={"ID":"3f53d421-a1b5-4495-9b35-d5e0805076e1","Type":"ContainerDied","Data":"57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682"} Oct 08 16:00:31 crc kubenswrapper[4945]: I1008 16:00:31.331843 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbtmn" event={"ID":"3f53d421-a1b5-4495-9b35-d5e0805076e1","Type":"ContainerStarted","Data":"7a746ac016e96dd99c24350efadd16f407cf9543b3b3b2d8c95c6da6397888f3"} Oct 08 16:00:33 crc kubenswrapper[4945]: I1008 16:00:33.351605 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbtmn" event={"ID":"3f53d421-a1b5-4495-9b35-d5e0805076e1","Type":"ContainerStarted","Data":"b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718"} Oct 08 16:00:34 crc kubenswrapper[4945]: I1008 16:00:34.362385 4945 generic.go:334] "Generic (PLEG): container finished" podID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerID="b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718" exitCode=0 Oct 08 16:00:34 crc kubenswrapper[4945]: I1008 16:00:34.362478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbtmn" event={"ID":"3f53d421-a1b5-4495-9b35-d5e0805076e1","Type":"ContainerDied","Data":"b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718"} Oct 08 16:00:36 crc kubenswrapper[4945]: I1008 16:00:36.382787 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbtmn" event={"ID":"3f53d421-a1b5-4495-9b35-d5e0805076e1","Type":"ContainerStarted","Data":"3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18"} Oct 08 16:00:36 crc kubenswrapper[4945]: I1008 16:00:36.404552 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rbtmn" podStartSLOduration=3.275317811 podStartE2EDuration="7.404529199s" podCreationTimestamp="2025-10-08 16:00:29 +0000 UTC" firstStartedPulling="2025-10-08 16:00:31.334065016 +0000 UTC m=+2840.687979917" lastFinishedPulling="2025-10-08 16:00:35.463276404 +0000 UTC m=+2844.817191305" observedRunningTime="2025-10-08 16:00:36.400919698 +0000 UTC m=+2845.754834599" watchObservedRunningTime="2025-10-08 16:00:36.404529199 +0000 UTC m=+2845.758444100" Oct 08 16:00:38 crc kubenswrapper[4945]: I1008 16:00:38.406285 4945 scope.go:117] "RemoveContainer" containerID="2e028ceec119da365ac3a511d19e8c0236278e93a94d7c9f0c81610de0295962" Oct 08 16:00:40 crc kubenswrapper[4945]: I1008 16:00:40.204687 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:40 crc kubenswrapper[4945]: I1008 16:00:40.205043 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:40 crc kubenswrapper[4945]: I1008 16:00:40.253087 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:40 crc kubenswrapper[4945]: I1008 16:00:40.478896 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:40 crc kubenswrapper[4945]: I1008 16:00:40.537416 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbtmn"] Oct 08 16:00:42 crc kubenswrapper[4945]: I1008 16:00:42.436917 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rbtmn" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="registry-server" containerID="cri-o://3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18" gracePeriod=2 Oct 08 16:00:42 crc kubenswrapper[4945]: I1008 16:00:42.913729 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.009855 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7zlg\" (UniqueName: \"kubernetes.io/projected/3f53d421-a1b5-4495-9b35-d5e0805076e1-kube-api-access-f7zlg\") pod \"3f53d421-a1b5-4495-9b35-d5e0805076e1\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.010021 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-utilities\") pod \"3f53d421-a1b5-4495-9b35-d5e0805076e1\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.010144 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-catalog-content\") pod \"3f53d421-a1b5-4495-9b35-d5e0805076e1\" (UID: \"3f53d421-a1b5-4495-9b35-d5e0805076e1\") " Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.011382 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-utilities" (OuterVolumeSpecName: "utilities") pod "3f53d421-a1b5-4495-9b35-d5e0805076e1" (UID: "3f53d421-a1b5-4495-9b35-d5e0805076e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.021386 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f53d421-a1b5-4495-9b35-d5e0805076e1-kube-api-access-f7zlg" (OuterVolumeSpecName: "kube-api-access-f7zlg") pod "3f53d421-a1b5-4495-9b35-d5e0805076e1" (UID: "3f53d421-a1b5-4495-9b35-d5e0805076e1"). InnerVolumeSpecName "kube-api-access-f7zlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.033623 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f53d421-a1b5-4495-9b35-d5e0805076e1" (UID: "3f53d421-a1b5-4495-9b35-d5e0805076e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.112565 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.112623 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7zlg\" (UniqueName: \"kubernetes.io/projected/3f53d421-a1b5-4495-9b35-d5e0805076e1-kube-api-access-f7zlg\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.112638 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f53d421-a1b5-4495-9b35-d5e0805076e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.451679 4945 generic.go:334] "Generic (PLEG): container finished" podID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerID="3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18" exitCode=0 Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.451730 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbtmn" event={"ID":"3f53d421-a1b5-4495-9b35-d5e0805076e1","Type":"ContainerDied","Data":"3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18"} Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.451762 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbtmn" event={"ID":"3f53d421-a1b5-4495-9b35-d5e0805076e1","Type":"ContainerDied","Data":"7a746ac016e96dd99c24350efadd16f407cf9543b3b3b2d8c95c6da6397888f3"} Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.451784 4945 scope.go:117] "RemoveContainer" containerID="3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.452018 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbtmn" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.482397 4945 scope.go:117] "RemoveContainer" containerID="b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.492369 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbtmn"] Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.501434 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbtmn"] Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.513755 4945 scope.go:117] "RemoveContainer" containerID="57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.585050 4945 scope.go:117] "RemoveContainer" containerID="3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18" Oct 08 16:00:43 crc kubenswrapper[4945]: E1008 16:00:43.585536 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18\": container with ID starting with 3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18 not found: ID does not exist" containerID="3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.585588 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18"} err="failed to get container status \"3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18\": rpc error: code = NotFound desc = could not find container \"3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18\": container with ID starting with 3a54470fa6032d51f3e6b1ed8e729befcd459b9a4ad45c3a36fd24f471c11f18 not found: ID does not exist" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.585619 4945 scope.go:117] "RemoveContainer" containerID="b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718" Oct 08 16:00:43 crc kubenswrapper[4945]: E1008 16:00:43.585976 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718\": container with ID starting with b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718 not found: ID does not exist" containerID="b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.586008 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718"} err="failed to get container status \"b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718\": rpc error: code = NotFound desc = could not find container \"b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718\": container with ID starting with b4ce55428208bd6ef97399af99aac3c6bdfa87d7720a538c89ef67602c788718 not found: ID does not exist" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.586028 4945 scope.go:117] "RemoveContainer" containerID="57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682" Oct 08 16:00:43 crc kubenswrapper[4945]: E1008 16:00:43.586436 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682\": container with ID starting with 57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682 not found: ID does not exist" containerID="57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682" Oct 08 16:00:43 crc kubenswrapper[4945]: I1008 16:00:43.586477 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682"} err="failed to get container status \"57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682\": rpc error: code = NotFound desc = could not find container \"57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682\": container with ID starting with 57b69f09a88a95cfda2a6d909e5bcd11d268998f2f43dac9156c4498f7333682 not found: ID does not exist" Oct 08 16:00:44 crc kubenswrapper[4945]: I1008 16:00:44.037305 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" path="/var/lib/kubelet/pods/3f53d421-a1b5-4495-9b35-d5e0805076e1/volumes" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.144627 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29332321-qc7kq"] Oct 08 16:01:00 crc kubenswrapper[4945]: E1008 16:01:00.145586 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="extract-content" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.145606 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="extract-content" Oct 08 16:01:00 crc kubenswrapper[4945]: E1008 16:01:00.145639 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="registry-server" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.145645 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="registry-server" Oct 08 16:01:00 crc kubenswrapper[4945]: E1008 16:01:00.145663 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="extract-utilities" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.145672 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="extract-utilities" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.145898 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f53d421-a1b5-4495-9b35-d5e0805076e1" containerName="registry-server" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.146607 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.157344 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332321-qc7kq"] Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.245880 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-combined-ca-bundle\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.245946 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzddw\" (UniqueName: \"kubernetes.io/projected/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-kube-api-access-lzddw\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.246082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-fernet-keys\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.246136 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-config-data\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.347439 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzddw\" (UniqueName: \"kubernetes.io/projected/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-kube-api-access-lzddw\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.347531 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-fernet-keys\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.347560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-config-data\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.347725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-combined-ca-bundle\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.354252 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-combined-ca-bundle\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.354717 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-fernet-keys\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.354819 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-config-data\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.363296 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzddw\" (UniqueName: \"kubernetes.io/projected/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-kube-api-access-lzddw\") pod \"keystone-cron-29332321-qc7kq\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.469188 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:00 crc kubenswrapper[4945]: I1008 16:01:00.921052 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332321-qc7kq"] Oct 08 16:01:01 crc kubenswrapper[4945]: I1008 16:01:01.632913 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332321-qc7kq" event={"ID":"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6","Type":"ContainerStarted","Data":"8e275cfff56162333856cd68c4645d7c7848c6586679955b64c7f872eaaa6d9e"} Oct 08 16:01:01 crc kubenswrapper[4945]: I1008 16:01:01.633245 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332321-qc7kq" event={"ID":"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6","Type":"ContainerStarted","Data":"3ad40e27c7b16653887336338d1fe2a4010be5fe1ff2eaaa9a7762b31d90cfde"} Oct 08 16:01:01 crc kubenswrapper[4945]: I1008 16:01:01.656311 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29332321-qc7kq" podStartSLOduration=1.6562869789999999 podStartE2EDuration="1.656286979s" podCreationTimestamp="2025-10-08 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 16:01:01.6519206 +0000 UTC m=+2871.005835501" watchObservedRunningTime="2025-10-08 16:01:01.656286979 +0000 UTC m=+2871.010201880" Oct 08 16:01:04 crc kubenswrapper[4945]: I1008 16:01:04.662301 4945 generic.go:334] "Generic (PLEG): container finished" podID="ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" containerID="8e275cfff56162333856cd68c4645d7c7848c6586679955b64c7f872eaaa6d9e" exitCode=0 Oct 08 16:01:04 crc kubenswrapper[4945]: I1008 16:01:04.662421 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332321-qc7kq" event={"ID":"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6","Type":"ContainerDied","Data":"8e275cfff56162333856cd68c4645d7c7848c6586679955b64c7f872eaaa6d9e"} Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.001939 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.171901 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-config-data\") pod \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.172172 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzddw\" (UniqueName: \"kubernetes.io/projected/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-kube-api-access-lzddw\") pod \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.172257 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-fernet-keys\") pod \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.172312 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-combined-ca-bundle\") pod \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\" (UID: \"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6\") " Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.177893 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" (UID: "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.178282 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-kube-api-access-lzddw" (OuterVolumeSpecName: "kube-api-access-lzddw") pod "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" (UID: "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6"). InnerVolumeSpecName "kube-api-access-lzddw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.206035 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" (UID: "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.246634 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-config-data" (OuterVolumeSpecName: "config-data") pod "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" (UID: "ca3e1262-6ad2-4df4-b86d-dc17fc2339c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.274600 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.274661 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.274674 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzddw\" (UniqueName: \"kubernetes.io/projected/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-kube-api-access-lzddw\") on node \"crc\" DevicePath \"\"" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.274710 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca3e1262-6ad2-4df4-b86d-dc17fc2339c6-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.682469 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332321-qc7kq" event={"ID":"ca3e1262-6ad2-4df4-b86d-dc17fc2339c6","Type":"ContainerDied","Data":"3ad40e27c7b16653887336338d1fe2a4010be5fe1ff2eaaa9a7762b31d90cfde"} Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.682509 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332321-qc7kq" Oct 08 16:01:06 crc kubenswrapper[4945]: I1008 16:01:06.682508 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ad40e27c7b16653887336338d1fe2a4010be5fe1ff2eaaa9a7762b31d90cfde" Oct 08 16:01:49 crc kubenswrapper[4945]: I1008 16:01:49.184419 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:01:49 crc kubenswrapper[4945]: I1008 16:01:49.185523 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:02:19 crc kubenswrapper[4945]: I1008 16:02:19.189946 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:02:19 crc kubenswrapper[4945]: I1008 16:02:19.191478 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:02:21 crc kubenswrapper[4945]: I1008 16:02:21.836420 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wbnrp"] Oct 08 16:02:21 crc kubenswrapper[4945]: E1008 16:02:21.837159 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" containerName="keystone-cron" Oct 08 16:02:21 crc kubenswrapper[4945]: I1008 16:02:21.837174 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" containerName="keystone-cron" Oct 08 16:02:21 crc kubenswrapper[4945]: I1008 16:02:21.837438 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca3e1262-6ad2-4df4-b86d-dc17fc2339c6" containerName="keystone-cron" Oct 08 16:02:21 crc kubenswrapper[4945]: I1008 16:02:21.839209 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:21 crc kubenswrapper[4945]: I1008 16:02:21.853017 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wbnrp"] Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.004433 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsntk\" (UniqueName: \"kubernetes.io/projected/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-kube-api-access-jsntk\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.004706 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-catalog-content\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.004737 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-utilities\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.106304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsntk\" (UniqueName: \"kubernetes.io/projected/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-kube-api-access-jsntk\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.106542 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-catalog-content\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.106580 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-utilities\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.107073 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-utilities\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.107211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-catalog-content\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.128193 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsntk\" (UniqueName: \"kubernetes.io/projected/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-kube-api-access-jsntk\") pod \"redhat-operators-wbnrp\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.160820 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:22 crc kubenswrapper[4945]: I1008 16:02:22.607496 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wbnrp"] Oct 08 16:02:23 crc kubenswrapper[4945]: I1008 16:02:23.434513 4945 generic.go:334] "Generic (PLEG): container finished" podID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerID="f369e2d18bb008538dfcdb3cf115bdbcddab9f8002e86fc74dde30e0774d4bb9" exitCode=0 Oct 08 16:02:23 crc kubenswrapper[4945]: I1008 16:02:23.434604 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbnrp" event={"ID":"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f","Type":"ContainerDied","Data":"f369e2d18bb008538dfcdb3cf115bdbcddab9f8002e86fc74dde30e0774d4bb9"} Oct 08 16:02:23 crc kubenswrapper[4945]: I1008 16:02:23.434900 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbnrp" event={"ID":"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f","Type":"ContainerStarted","Data":"c3996e8ea87a0e06971419debe6b5125fa5f8391ed14fb9441467b89bcfe41dd"} Oct 08 16:02:25 crc kubenswrapper[4945]: I1008 16:02:25.458687 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbnrp" event={"ID":"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f","Type":"ContainerStarted","Data":"9334b633912154d5a07675395e572e3e4f7a03ee0b0f11ce71d5126d5f63dcf8"} Oct 08 16:02:26 crc kubenswrapper[4945]: I1008 16:02:26.474306 4945 generic.go:334] "Generic (PLEG): container finished" podID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerID="9334b633912154d5a07675395e572e3e4f7a03ee0b0f11ce71d5126d5f63dcf8" exitCode=0 Oct 08 16:02:26 crc kubenswrapper[4945]: I1008 16:02:26.474376 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbnrp" event={"ID":"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f","Type":"ContainerDied","Data":"9334b633912154d5a07675395e572e3e4f7a03ee0b0f11ce71d5126d5f63dcf8"} Oct 08 16:02:28 crc kubenswrapper[4945]: I1008 16:02:28.496973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbnrp" event={"ID":"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f","Type":"ContainerStarted","Data":"25291cef51dbdbe9f63d8e75a049c6415369c172fc97380ccc6035814d3a7927"} Oct 08 16:02:28 crc kubenswrapper[4945]: I1008 16:02:28.514087 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wbnrp" podStartSLOduration=3.209445245 podStartE2EDuration="7.514047753s" podCreationTimestamp="2025-10-08 16:02:21 +0000 UTC" firstStartedPulling="2025-10-08 16:02:23.436750568 +0000 UTC m=+2952.790665499" lastFinishedPulling="2025-10-08 16:02:27.741353106 +0000 UTC m=+2957.095268007" observedRunningTime="2025-10-08 16:02:28.512591837 +0000 UTC m=+2957.866506748" watchObservedRunningTime="2025-10-08 16:02:28.514047753 +0000 UTC m=+2957.867962664" Oct 08 16:02:32 crc kubenswrapper[4945]: I1008 16:02:32.161104 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:32 crc kubenswrapper[4945]: I1008 16:02:32.161507 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:32 crc kubenswrapper[4945]: I1008 16:02:32.220406 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:32 crc kubenswrapper[4945]: I1008 16:02:32.589571 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:32 crc kubenswrapper[4945]: I1008 16:02:32.651748 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wbnrp"] Oct 08 16:02:34 crc kubenswrapper[4945]: I1008 16:02:34.554760 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wbnrp" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="registry-server" containerID="cri-o://25291cef51dbdbe9f63d8e75a049c6415369c172fc97380ccc6035814d3a7927" gracePeriod=2 Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.566068 4945 generic.go:334] "Generic (PLEG): container finished" podID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerID="25291cef51dbdbe9f63d8e75a049c6415369c172fc97380ccc6035814d3a7927" exitCode=0 Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.566159 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbnrp" event={"ID":"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f","Type":"ContainerDied","Data":"25291cef51dbdbe9f63d8e75a049c6415369c172fc97380ccc6035814d3a7927"} Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.566417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbnrp" event={"ID":"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f","Type":"ContainerDied","Data":"c3996e8ea87a0e06971419debe6b5125fa5f8391ed14fb9441467b89bcfe41dd"} Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.566441 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3996e8ea87a0e06971419debe6b5125fa5f8391ed14fb9441467b89bcfe41dd" Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.623851 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.787215 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-catalog-content\") pod \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.787349 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsntk\" (UniqueName: \"kubernetes.io/projected/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-kube-api-access-jsntk\") pod \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.787524 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-utilities\") pod \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\" (UID: \"38d1bcdc-fac1-4986-b7c1-683d7d91ce2f\") " Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.788563 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-utilities" (OuterVolumeSpecName: "utilities") pod "38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" (UID: "38d1bcdc-fac1-4986-b7c1-683d7d91ce2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.793660 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-kube-api-access-jsntk" (OuterVolumeSpecName: "kube-api-access-jsntk") pod "38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" (UID: "38d1bcdc-fac1-4986-b7c1-683d7d91ce2f"). InnerVolumeSpecName "kube-api-access-jsntk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.879285 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" (UID: "38d1bcdc-fac1-4986-b7c1-683d7d91ce2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.890324 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsntk\" (UniqueName: \"kubernetes.io/projected/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-kube-api-access-jsntk\") on node \"crc\" DevicePath \"\"" Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.890368 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:02:35 crc kubenswrapper[4945]: I1008 16:02:35.890384 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:02:36 crc kubenswrapper[4945]: I1008 16:02:36.576393 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbnrp" Oct 08 16:02:36 crc kubenswrapper[4945]: I1008 16:02:36.609783 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wbnrp"] Oct 08 16:02:36 crc kubenswrapper[4945]: I1008 16:02:36.622593 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wbnrp"] Oct 08 16:02:38 crc kubenswrapper[4945]: I1008 16:02:38.036224 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" path="/var/lib/kubelet/pods/38d1bcdc-fac1-4986-b7c1-683d7d91ce2f/volumes" Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.183711 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.184229 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.184269 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.184893 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.184943 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" gracePeriod=600 Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.710253 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" exitCode=0 Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.710298 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609"} Oct 08 16:02:49 crc kubenswrapper[4945]: I1008 16:02:49.710354 4945 scope.go:117] "RemoveContainer" containerID="cbd03bbb5effbe7203ee78f496768b2e3ade6cfa776d6ff5e8810e49735e8777" Oct 08 16:02:49 crc kubenswrapper[4945]: E1008 16:02:49.923919 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:02:50 crc kubenswrapper[4945]: I1008 16:02:50.726937 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:02:50 crc kubenswrapper[4945]: E1008 16:02:50.727455 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:03:06 crc kubenswrapper[4945]: I1008 16:03:06.036572 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:03:06 crc kubenswrapper[4945]: E1008 16:03:06.039018 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:03:20 crc kubenswrapper[4945]: I1008 16:03:20.024706 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:03:20 crc kubenswrapper[4945]: E1008 16:03:20.025590 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:03:35 crc kubenswrapper[4945]: I1008 16:03:35.024844 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:03:35 crc kubenswrapper[4945]: E1008 16:03:35.025879 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:03:43 crc kubenswrapper[4945]: I1008 16:03:43.294808 4945 generic.go:334] "Generic (PLEG): container finished" podID="8b121934-fc40-487d-91fe-cdc91272f2dd" containerID="a1ca5691ca7108e29e13a477a810732c62df29f5690b442f7c1a1084d7267cf5" exitCode=0 Oct 08 16:03:43 crc kubenswrapper[4945]: I1008 16:03:43.294860 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" event={"ID":"8b121934-fc40-487d-91fe-cdc91272f2dd","Type":"ContainerDied","Data":"a1ca5691ca7108e29e13a477a810732c62df29f5690b442f7c1a1084d7267cf5"} Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.713646 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.842730 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-inventory\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.842825 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-0\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.842869 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-0\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.842922 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-combined-ca-bundle\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.842995 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr6jr\" (UniqueName: \"kubernetes.io/projected/8b121934-fc40-487d-91fe-cdc91272f2dd-kube-api-access-nr6jr\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.843068 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-1\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.843148 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-extra-config-0\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.843213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-ssh-key\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.843237 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-1\") pod \"8b121934-fc40-487d-91fe-cdc91272f2dd\" (UID: \"8b121934-fc40-487d-91fe-cdc91272f2dd\") " Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.849758 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b121934-fc40-487d-91fe-cdc91272f2dd-kube-api-access-nr6jr" (OuterVolumeSpecName: "kube-api-access-nr6jr") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "kube-api-access-nr6jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.857327 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.891675 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.896730 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.897537 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.909510 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.912208 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-inventory" (OuterVolumeSpecName: "inventory") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.918405 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.922660 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "8b121934-fc40-487d-91fe-cdc91272f2dd" (UID: "8b121934-fc40-487d-91fe-cdc91272f2dd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945428 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr6jr\" (UniqueName: \"kubernetes.io/projected/8b121934-fc40-487d-91fe-cdc91272f2dd-kube-api-access-nr6jr\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945463 4945 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945476 4945 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945493 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945505 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945518 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945545 4945 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945557 4945 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:44 crc kubenswrapper[4945]: I1008 16:03:44.945569 4945 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b121934-fc40-487d-91fe-cdc91272f2dd-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.318931 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" event={"ID":"8b121934-fc40-487d-91fe-cdc91272f2dd","Type":"ContainerDied","Data":"96788fa4c49e2863b856ab352ac1a386b38cc6c844a8cbf941cfcac822c5f220"} Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.319310 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96788fa4c49e2863b856ab352ac1a386b38cc6c844a8cbf941cfcac822c5f220" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.319000 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mpr75" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.463965 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5"] Oct 08 16:03:45 crc kubenswrapper[4945]: E1008 16:03:45.464455 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b121934-fc40-487d-91fe-cdc91272f2dd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.464480 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b121934-fc40-487d-91fe-cdc91272f2dd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 16:03:45 crc kubenswrapper[4945]: E1008 16:03:45.464521 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="registry-server" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.464533 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="registry-server" Oct 08 16:03:45 crc kubenswrapper[4945]: E1008 16:03:45.464561 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="extract-content" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.464569 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="extract-content" Oct 08 16:03:45 crc kubenswrapper[4945]: E1008 16:03:45.464593 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="extract-utilities" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.464601 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="extract-utilities" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.464846 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="38d1bcdc-fac1-4986-b7c1-683d7d91ce2f" containerName="registry-server" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.464872 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b121934-fc40-487d-91fe-cdc91272f2dd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.465687 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.467839 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.468030 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fsbhh" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.473929 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.473929 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.475554 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.482598 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5"] Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.556814 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.556870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.557014 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.557145 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.557249 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42z74\" (UniqueName: \"kubernetes.io/projected/d7a976cb-af44-4232-8415-da43e89bff8a-kube-api-access-42z74\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.557286 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.557311 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.658680 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42z74\" (UniqueName: \"kubernetes.io/projected/d7a976cb-af44-4232-8415-da43e89bff8a-kube-api-access-42z74\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.658752 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.658776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.658816 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.658841 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.658940 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.659020 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.662949 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.663542 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.664604 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.664695 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.665051 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.665738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.675802 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42z74\" (UniqueName: \"kubernetes.io/projected/d7a976cb-af44-4232-8415-da43e89bff8a-kube-api-access-42z74\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v89t5\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:45 crc kubenswrapper[4945]: I1008 16:03:45.795838 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:03:46 crc kubenswrapper[4945]: I1008 16:03:46.325718 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5"] Oct 08 16:03:47 crc kubenswrapper[4945]: I1008 16:03:47.024946 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:03:47 crc kubenswrapper[4945]: E1008 16:03:47.025521 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:03:47 crc kubenswrapper[4945]: I1008 16:03:47.341021 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" event={"ID":"d7a976cb-af44-4232-8415-da43e89bff8a","Type":"ContainerStarted","Data":"d3abf8a2ce023624551f3ceb1f163547abbfbf3b45cdb2bf2427456bffbaf3bc"} Oct 08 16:03:48 crc kubenswrapper[4945]: I1008 16:03:48.350996 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" event={"ID":"d7a976cb-af44-4232-8415-da43e89bff8a","Type":"ContainerStarted","Data":"d8c4bc71c3d4070348a387d7a273ba0d5ee24be31bc57a9c77590d765f4fc666"} Oct 08 16:03:48 crc kubenswrapper[4945]: I1008 16:03:48.377138 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" podStartSLOduration=2.54215628 podStartE2EDuration="3.377098253s" podCreationTimestamp="2025-10-08 16:03:45 +0000 UTC" firstStartedPulling="2025-10-08 16:03:46.332871686 +0000 UTC m=+3035.686786587" lastFinishedPulling="2025-10-08 16:03:47.167813659 +0000 UTC m=+3036.521728560" observedRunningTime="2025-10-08 16:03:48.367775929 +0000 UTC m=+3037.721690840" watchObservedRunningTime="2025-10-08 16:03:48.377098253 +0000 UTC m=+3037.731013164" Oct 08 16:04:02 crc kubenswrapper[4945]: I1008 16:04:02.030143 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:04:02 crc kubenswrapper[4945]: E1008 16:04:02.030899 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:04:13 crc kubenswrapper[4945]: I1008 16:04:13.024211 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:04:13 crc kubenswrapper[4945]: E1008 16:04:13.024910 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:04:26 crc kubenswrapper[4945]: I1008 16:04:26.024984 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:04:26 crc kubenswrapper[4945]: E1008 16:04:26.026252 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:04:39 crc kubenswrapper[4945]: I1008 16:04:39.025029 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:04:39 crc kubenswrapper[4945]: E1008 16:04:39.025910 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:04:51 crc kubenswrapper[4945]: I1008 16:04:51.025321 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:04:51 crc kubenswrapper[4945]: E1008 16:04:51.026267 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:05:05 crc kubenswrapper[4945]: I1008 16:05:05.024678 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:05:05 crc kubenswrapper[4945]: E1008 16:05:05.025630 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:05:16 crc kubenswrapper[4945]: I1008 16:05:16.025269 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:05:16 crc kubenswrapper[4945]: E1008 16:05:16.026402 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:05:30 crc kubenswrapper[4945]: I1008 16:05:30.024526 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:05:30 crc kubenswrapper[4945]: E1008 16:05:30.025202 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:05:45 crc kubenswrapper[4945]: I1008 16:05:45.024408 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:05:45 crc kubenswrapper[4945]: E1008 16:05:45.025497 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:06:00 crc kubenswrapper[4945]: I1008 16:06:00.024829 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:06:00 crc kubenswrapper[4945]: E1008 16:06:00.025735 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:06:08 crc kubenswrapper[4945]: I1008 16:06:08.751695 4945 generic.go:334] "Generic (PLEG): container finished" podID="d7a976cb-af44-4232-8415-da43e89bff8a" containerID="d8c4bc71c3d4070348a387d7a273ba0d5ee24be31bc57a9c77590d765f4fc666" exitCode=0 Oct 08 16:06:08 crc kubenswrapper[4945]: I1008 16:06:08.751797 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" event={"ID":"d7a976cb-af44-4232-8415-da43e89bff8a","Type":"ContainerDied","Data":"d8c4bc71c3d4070348a387d7a273ba0d5ee24be31bc57a9c77590d765f4fc666"} Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.263018 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.329849 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-telemetry-combined-ca-bundle\") pod \"d7a976cb-af44-4232-8415-da43e89bff8a\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.329914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-1\") pod \"d7a976cb-af44-4232-8415-da43e89bff8a\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.330047 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-2\") pod \"d7a976cb-af44-4232-8415-da43e89bff8a\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.330142 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ssh-key\") pod \"d7a976cb-af44-4232-8415-da43e89bff8a\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.330222 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-0\") pod \"d7a976cb-af44-4232-8415-da43e89bff8a\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.330247 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-inventory\") pod \"d7a976cb-af44-4232-8415-da43e89bff8a\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.330283 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42z74\" (UniqueName: \"kubernetes.io/projected/d7a976cb-af44-4232-8415-da43e89bff8a-kube-api-access-42z74\") pod \"d7a976cb-af44-4232-8415-da43e89bff8a\" (UID: \"d7a976cb-af44-4232-8415-da43e89bff8a\") " Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.370296 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a976cb-af44-4232-8415-da43e89bff8a-kube-api-access-42z74" (OuterVolumeSpecName: "kube-api-access-42z74") pod "d7a976cb-af44-4232-8415-da43e89bff8a" (UID: "d7a976cb-af44-4232-8415-da43e89bff8a"). InnerVolumeSpecName "kube-api-access-42z74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.375801 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d7a976cb-af44-4232-8415-da43e89bff8a" (UID: "d7a976cb-af44-4232-8415-da43e89bff8a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.437529 4945 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.437565 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42z74\" (UniqueName: \"kubernetes.io/projected/d7a976cb-af44-4232-8415-da43e89bff8a-kube-api-access-42z74\") on node \"crc\" DevicePath \"\"" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.450087 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d7a976cb-af44-4232-8415-da43e89bff8a" (UID: "d7a976cb-af44-4232-8415-da43e89bff8a"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.452672 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d7a976cb-af44-4232-8415-da43e89bff8a" (UID: "d7a976cb-af44-4232-8415-da43e89bff8a"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.453093 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-inventory" (OuterVolumeSpecName: "inventory") pod "d7a976cb-af44-4232-8415-da43e89bff8a" (UID: "d7a976cb-af44-4232-8415-da43e89bff8a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.455403 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d7a976cb-af44-4232-8415-da43e89bff8a" (UID: "d7a976cb-af44-4232-8415-da43e89bff8a"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.473025 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d7a976cb-af44-4232-8415-da43e89bff8a" (UID: "d7a976cb-af44-4232-8415-da43e89bff8a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.539082 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.539147 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.539170 4945 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.539186 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.539199 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7a976cb-af44-4232-8415-da43e89bff8a-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.778956 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" event={"ID":"d7a976cb-af44-4232-8415-da43e89bff8a","Type":"ContainerDied","Data":"d3abf8a2ce023624551f3ceb1f163547abbfbf3b45cdb2bf2427456bffbaf3bc"} Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.779005 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3abf8a2ce023624551f3ceb1f163547abbfbf3b45cdb2bf2427456bffbaf3bc" Oct 08 16:06:10 crc kubenswrapper[4945]: I1008 16:06:10.779081 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v89t5" Oct 08 16:06:12 crc kubenswrapper[4945]: I1008 16:06:12.051697 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:06:12 crc kubenswrapper[4945]: E1008 16:06:12.053397 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:06:23 crc kubenswrapper[4945]: I1008 16:06:23.024162 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:06:23 crc kubenswrapper[4945]: E1008 16:06:23.025010 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:06:39 crc kubenswrapper[4945]: I1008 16:06:39.025020 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:06:39 crc kubenswrapper[4945]: E1008 16:06:39.026318 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.195791 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 08 16:06:45 crc kubenswrapper[4945]: E1008 16:06:45.196708 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a976cb-af44-4232-8415-da43e89bff8a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.196725 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a976cb-af44-4232-8415-da43e89bff8a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.196920 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a976cb-af44-4232-8415-da43e89bff8a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.197968 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.200719 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.221287 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.298954 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.304720 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.331883 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.332845 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-lib-modules\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.332919 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-scripts\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.332963 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.332996 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-config-data\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333126 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333172 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lczh\" (UniqueName: \"kubernetes.io/projected/a2eda97e-ccd7-42e8-bca5-26402069c250-kube-api-access-9lczh\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333254 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-config-data-custom\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333464 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333490 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333532 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-dev\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333838 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-nvme\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333870 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-sys\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.333899 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-run\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.338842 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-2-config-data" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.353186 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-0"] Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.355178 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.357921 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-config-data" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.362876 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435197 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435239 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435269 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-lib-modules\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435472 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-scripts\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435498 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435532 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435552 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435579 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-config-data\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435662 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435688 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435720 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435758 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z92q\" (UniqueName: \"kubernetes.io/projected/e2f8bf86-ab2b-428a-9e89-2d71091927ef-kube-api-access-8z92q\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lczh\" (UniqueName: \"kubernetes.io/projected/a2eda97e-ccd7-42e8-bca5-26402069c250-kube-api-access-9lczh\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435797 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97zr2\" (UniqueName: \"kubernetes.io/projected/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-kube-api-access-97zr2\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435837 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435902 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435917 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435944 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.435994 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436023 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436048 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-dev\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436073 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436127 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-config-data-custom\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436166 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436215 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436232 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436247 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436267 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436284 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-sys\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436315 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-dev\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436446 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-lib-modules\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436615 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436691 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-dev\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.436816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437231 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437295 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437422 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-run\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437533 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437565 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-nvme\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437581 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-sys\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437612 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437629 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-run\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437721 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-run\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.437800 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-sys\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.438245 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-etc-nvme\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.438359 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a2eda97e-ccd7-42e8-bca5-26402069c250-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.442877 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-config-data-custom\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.443826 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-config-data\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.445637 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-scripts\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.446182 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2eda97e-ccd7-42e8-bca5-26402069c250-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.464356 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lczh\" (UniqueName: \"kubernetes.io/projected/a2eda97e-ccd7-42e8-bca5-26402069c250-kube-api-access-9lczh\") pod \"cinder-backup-0\" (UID: \"a2eda97e-ccd7-42e8-bca5-26402069c250\") " pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.520707 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.542810 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.542897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.542940 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z92q\" (UniqueName: \"kubernetes.io/projected/e2f8bf86-ab2b-428a-9e89-2d71091927ef-kube-api-access-8z92q\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.542958 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97zr2\" (UniqueName: \"kubernetes.io/projected/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-kube-api-access-97zr2\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.542964 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543020 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.542977 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543090 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543205 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543237 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543260 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543294 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543302 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543330 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-dev\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543354 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543369 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-dev\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543429 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543532 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543555 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543588 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-sys\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543609 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543648 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543673 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543696 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543712 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-run\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543741 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543755 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543772 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543796 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543817 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543831 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543865 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543880 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543896 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543911 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543920 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-run\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-sys\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.543937 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.544018 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.544050 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.544095 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.544292 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.544626 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.544684 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.545039 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e2f8bf86-ab2b-428a-9e89-2d71091927ef-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.548875 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.550030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.550506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.551494 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.560071 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.561354 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.561610 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.563892 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f8bf86-ab2b-428a-9e89-2d71091927ef-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.566721 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z92q\" (UniqueName: \"kubernetes.io/projected/e2f8bf86-ab2b-428a-9e89-2d71091927ef-kube-api-access-8z92q\") pod \"cinder-volume-nfs-2-0\" (UID: \"e2f8bf86-ab2b-428a-9e89-2d71091927ef\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.568595 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97zr2\" (UniqueName: \"kubernetes.io/projected/1c0d0627-7d92-4bd7-8f53-3cd9117aec3b-kube-api-access-97zr2\") pod \"cinder-volume-nfs-0\" (UID: \"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.678009 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:45 crc kubenswrapper[4945]: I1008 16:06:45.690155 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:46 crc kubenswrapper[4945]: I1008 16:06:46.089929 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 08 16:06:46 crc kubenswrapper[4945]: I1008 16:06:46.097829 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 16:06:46 crc kubenswrapper[4945]: W1008 16:06:46.859984 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2f8bf86_ab2b_428a_9e89_2d71091927ef.slice/crio-729cc7a9f728f39edc44b4161903609f326129fc697bd3916aef67af745b705f WatchSource:0}: Error finding container 729cc7a9f728f39edc44b4161903609f326129fc697bd3916aef67af745b705f: Status 404 returned error can't find the container with id 729cc7a9f728f39edc44b4161903609f326129fc697bd3916aef67af745b705f Oct 08 16:06:46 crc kubenswrapper[4945]: I1008 16:06:46.864810 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Oct 08 16:06:46 crc kubenswrapper[4945]: I1008 16:06:46.962237 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Oct 08 16:06:47 crc kubenswrapper[4945]: W1008 16:06:47.033208 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c0d0627_7d92_4bd7_8f53_3cd9117aec3b.slice/crio-6bef726a2ad8795da0e2c70977d0193870d1dc0705020bcf36da42c80de0e61d WatchSource:0}: Error finding container 6bef726a2ad8795da0e2c70977d0193870d1dc0705020bcf36da42c80de0e61d: Status 404 returned error can't find the container with id 6bef726a2ad8795da0e2c70977d0193870d1dc0705020bcf36da42c80de0e61d Oct 08 16:06:47 crc kubenswrapper[4945]: I1008 16:06:47.118131 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"e2f8bf86-ab2b-428a-9e89-2d71091927ef","Type":"ContainerStarted","Data":"729cc7a9f728f39edc44b4161903609f326129fc697bd3916aef67af745b705f"} Oct 08 16:06:47 crc kubenswrapper[4945]: I1008 16:06:47.120431 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"a2eda97e-ccd7-42e8-bca5-26402069c250","Type":"ContainerStarted","Data":"1c57e2f5de516e27fe9833694137b0b28a39746966770ddcb31502763266b84c"} Oct 08 16:06:47 crc kubenswrapper[4945]: I1008 16:06:47.120531 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"a2eda97e-ccd7-42e8-bca5-26402069c250","Type":"ContainerStarted","Data":"97a12517861f37eb5830fbef25e466fe28675543b1351dbe831def5813b19b91"} Oct 08 16:06:47 crc kubenswrapper[4945]: I1008 16:06:47.120546 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"a2eda97e-ccd7-42e8-bca5-26402069c250","Type":"ContainerStarted","Data":"2766fd6638d48af64948b9e14d400af7d9f92b4b5a0776571da4fc3390ea32da"} Oct 08 16:06:47 crc kubenswrapper[4945]: I1008 16:06:47.121710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b","Type":"ContainerStarted","Data":"6bef726a2ad8795da0e2c70977d0193870d1dc0705020bcf36da42c80de0e61d"} Oct 08 16:06:47 crc kubenswrapper[4945]: I1008 16:06:47.146959 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=1.93026294 podStartE2EDuration="2.146934497s" podCreationTimestamp="2025-10-08 16:06:45 +0000 UTC" firstStartedPulling="2025-10-08 16:06:46.097553054 +0000 UTC m=+3215.451467955" lastFinishedPulling="2025-10-08 16:06:46.314224611 +0000 UTC m=+3215.668139512" observedRunningTime="2025-10-08 16:06:47.144925323 +0000 UTC m=+3216.498840224" watchObservedRunningTime="2025-10-08 16:06:47.146934497 +0000 UTC m=+3216.500849408" Oct 08 16:06:48 crc kubenswrapper[4945]: I1008 16:06:48.138142 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"e2f8bf86-ab2b-428a-9e89-2d71091927ef","Type":"ContainerStarted","Data":"75382b98c29c9182d327abc96bcc54b7417123e9b5a8bf0c1008ef2c2d34e681"} Oct 08 16:06:48 crc kubenswrapper[4945]: I1008 16:06:48.140336 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"e2f8bf86-ab2b-428a-9e89-2d71091927ef","Type":"ContainerStarted","Data":"9632ae7a7909f9683565a5a91fdaef8ee986c01759f784af4c189bac6ad74843"} Oct 08 16:06:48 crc kubenswrapper[4945]: I1008 16:06:48.143385 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b","Type":"ContainerStarted","Data":"1ef9d8c1e2cd08f600f3d6a2e3c50e4190859a1ff76f0184b0ea8e6aa7225d94"} Oct 08 16:06:48 crc kubenswrapper[4945]: I1008 16:06:48.143463 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"1c0d0627-7d92-4bd7-8f53-3cd9117aec3b","Type":"ContainerStarted","Data":"03bc69656b0298d7ced9d99c35fee00997edf7d005861c3d9de528fb83662bd3"} Oct 08 16:06:48 crc kubenswrapper[4945]: I1008 16:06:48.178565 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-2-0" podStartSLOduration=2.934285591 podStartE2EDuration="3.178543894s" podCreationTimestamp="2025-10-08 16:06:45 +0000 UTC" firstStartedPulling="2025-10-08 16:06:46.862137497 +0000 UTC m=+3216.216052408" lastFinishedPulling="2025-10-08 16:06:47.10639581 +0000 UTC m=+3216.460310711" observedRunningTime="2025-10-08 16:06:48.174195577 +0000 UTC m=+3217.528110478" watchObservedRunningTime="2025-10-08 16:06:48.178543894 +0000 UTC m=+3217.532458805" Oct 08 16:06:48 crc kubenswrapper[4945]: I1008 16:06:48.210292 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-0" podStartSLOduration=3.136248948 podStartE2EDuration="3.210263914s" podCreationTimestamp="2025-10-08 16:06:45 +0000 UTC" firstStartedPulling="2025-10-08 16:06:47.035216169 +0000 UTC m=+3216.389131070" lastFinishedPulling="2025-10-08 16:06:47.109231135 +0000 UTC m=+3216.463146036" observedRunningTime="2025-10-08 16:06:48.200489845 +0000 UTC m=+3217.554404766" watchObservedRunningTime="2025-10-08 16:06:48.210263914 +0000 UTC m=+3217.564178845" Oct 08 16:06:50 crc kubenswrapper[4945]: I1008 16:06:50.522756 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 08 16:06:50 crc kubenswrapper[4945]: I1008 16:06:50.678732 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:50 crc kubenswrapper[4945]: I1008 16:06:50.690735 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-0" Oct 08 16:06:54 crc kubenswrapper[4945]: I1008 16:06:54.028162 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:06:54 crc kubenswrapper[4945]: E1008 16:06:54.030678 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:06:55 crc kubenswrapper[4945]: I1008 16:06:55.702606 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 08 16:06:55 crc kubenswrapper[4945]: I1008 16:06:55.888888 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-2-0" Oct 08 16:06:55 crc kubenswrapper[4945]: I1008 16:06:55.973429 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-0" Oct 08 16:07:05 crc kubenswrapper[4945]: I1008 16:07:05.023982 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:07:05 crc kubenswrapper[4945]: E1008 16:07:05.024849 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:07:16 crc kubenswrapper[4945]: I1008 16:07:16.024605 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:07:16 crc kubenswrapper[4945]: E1008 16:07:16.025715 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:07:30 crc kubenswrapper[4945]: I1008 16:07:30.024962 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:07:30 crc kubenswrapper[4945]: E1008 16:07:30.025724 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:07:41 crc kubenswrapper[4945]: I1008 16:07:41.025500 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:07:41 crc kubenswrapper[4945]: E1008 16:07:41.026590 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:07:48 crc kubenswrapper[4945]: I1008 16:07:48.650254 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 16:07:48 crc kubenswrapper[4945]: I1008 16:07:48.651054 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="prometheus" containerID="cri-o://e93d65b01947105be03f6d85d3cb05be4b1a2bd6929ee804b1afffa14db7fd1f" gracePeriod=600 Oct 08 16:07:48 crc kubenswrapper[4945]: I1008 16:07:48.651131 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="thanos-sidecar" containerID="cri-o://224dbcb223e227fbba9cd303b145f9619c8aa4605a720c0d2ebbb675ceb3f2c2" gracePeriod=600 Oct 08 16:07:48 crc kubenswrapper[4945]: I1008 16:07:48.651185 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="config-reloader" containerID="cri-o://f1f98955cf4183a5a254adcd49d46fb488561128e63fdb23c2e396868cbf0d1c" gracePeriod=600 Oct 08 16:07:49 crc kubenswrapper[4945]: I1008 16:07:49.772566 4945 generic.go:334] "Generic (PLEG): container finished" podID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerID="224dbcb223e227fbba9cd303b145f9619c8aa4605a720c0d2ebbb675ceb3f2c2" exitCode=0 Oct 08 16:07:49 crc kubenswrapper[4945]: I1008 16:07:49.773112 4945 generic.go:334] "Generic (PLEG): container finished" podID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerID="f1f98955cf4183a5a254adcd49d46fb488561128e63fdb23c2e396868cbf0d1c" exitCode=0 Oct 08 16:07:49 crc kubenswrapper[4945]: I1008 16:07:49.773135 4945 generic.go:334] "Generic (PLEG): container finished" podID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerID="e93d65b01947105be03f6d85d3cb05be4b1a2bd6929ee804b1afffa14db7fd1f" exitCode=0 Oct 08 16:07:49 crc kubenswrapper[4945]: I1008 16:07:49.773152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerDied","Data":"224dbcb223e227fbba9cd303b145f9619c8aa4605a720c0d2ebbb675ceb3f2c2"} Oct 08 16:07:49 crc kubenswrapper[4945]: I1008 16:07:49.773180 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerDied","Data":"f1f98955cf4183a5a254adcd49d46fb488561128e63fdb23c2e396868cbf0d1c"} Oct 08 16:07:49 crc kubenswrapper[4945]: I1008 16:07:49.773189 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerDied","Data":"e93d65b01947105be03f6d85d3cb05be4b1a2bd6929ee804b1afffa14db7fd1f"} Oct 08 16:07:49 crc kubenswrapper[4945]: I1008 16:07:49.963764 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035263 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-tls-assets\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035309 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-secret-combined-ca-bundle\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035342 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035371 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035411 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config-out\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035613 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-thanos-prometheus-http-client-file\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035675 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035707 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-prometheus-metric-storage-rulefiles-0\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.035767 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4fwf\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-kube-api-access-q4fwf\") pod \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\" (UID: \"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5\") " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.040389 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.042832 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.042907 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.044950 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.048362 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.048809 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config-out" (OuterVolumeSpecName: "config-out") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.049321 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.049372 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-kube-api-access-q4fwf" (OuterVolumeSpecName: "kube-api-access-q4fwf") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "kube-api-access-q4fwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.051745 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config" (OuterVolumeSpecName: "config") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.083065 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139149 4945 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139522 4945 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139542 4945 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139580 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139595 4945 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-config-out\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139845 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") on node \"crc\" " Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139861 4945 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139876 4945 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139909 4945 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.139923 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4fwf\" (UniqueName: \"kubernetes.io/projected/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-kube-api-access-q4fwf\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.175022 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config" (OuterVolumeSpecName: "web-config") pod "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" (UID: "e540f1d4-87f5-4ddf-a9a0-d8141141b9d5"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.188546 4945 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.190335 4945 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e") on node "crc" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.241584 4945 reconciler_common.go:293] "Volume detached for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.241635 4945 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5-web-config\") on node \"crc\" DevicePath \"\"" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.785284 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e540f1d4-87f5-4ddf-a9a0-d8141141b9d5","Type":"ContainerDied","Data":"e02aa4b57adaaf2ba323f3d8cd27a5c3b0789317399f84f197fe9cac0c4ca889"} Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.785346 4945 scope.go:117] "RemoveContainer" containerID="224dbcb223e227fbba9cd303b145f9619c8aa4605a720c0d2ebbb675ceb3f2c2" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.785394 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.835320 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.837388 4945 scope.go:117] "RemoveContainer" containerID="f1f98955cf4183a5a254adcd49d46fb488561128e63fdb23c2e396868cbf0d1c" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.849107 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.872142 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 16:07:50 crc kubenswrapper[4945]: E1008 16:07:50.872658 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="init-config-reloader" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.872681 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="init-config-reloader" Oct 08 16:07:50 crc kubenswrapper[4945]: E1008 16:07:50.872703 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="config-reloader" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.872710 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="config-reloader" Oct 08 16:07:50 crc kubenswrapper[4945]: E1008 16:07:50.872737 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="thanos-sidecar" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.872744 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="thanos-sidecar" Oct 08 16:07:50 crc kubenswrapper[4945]: E1008 16:07:50.872754 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="prometheus" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.872764 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="prometheus" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.872983 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="config-reloader" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.873013 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="thanos-sidecar" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.873043 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" containerName="prometheus" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.876290 4945 scope.go:117] "RemoveContainer" containerID="e93d65b01947105be03f6d85d3cb05be4b1a2bd6929ee804b1afffa14db7fd1f" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.879899 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.883749 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.886506 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.887073 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.887756 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-qmfdz" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.888295 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.893201 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.914180 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.932715 4945 scope.go:117] "RemoveContainer" containerID="81c5bd76ee389f38023724b07d0a44bc55ee8f44508d56fe54a7e0e51d2e56f8" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957457 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957585 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957619 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957635 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957658 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957720 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957742 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957764 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957799 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957830 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mrw8\" (UniqueName: \"kubernetes.io/projected/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-kube-api-access-9mrw8\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:50 crc kubenswrapper[4945]: I1008 16:07:50.957875 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.059887 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060394 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060437 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060462 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060492 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060571 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060610 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060736 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mrw8\" (UniqueName: \"kubernetes.io/projected/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-kube-api-access-9mrw8\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.060983 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.063493 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.063540 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9dcd0c273e01834b6d6b9a81fac4d1c0f2ca078125a1dee333b69bb46fb7d5e3/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.065727 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.065783 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.065881 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.066383 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.067850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.068136 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.068142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.073262 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.083519 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mrw8\" (UniqueName: \"kubernetes.io/projected/9ce8e6f8-1567-4bd4-8f2c-ada70838ec92-kube-api-access-9mrw8\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.105682 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad65107f-6e91-4bcc-b89c-4cdb7683610e\") pod \"prometheus-metric-storage-0\" (UID: \"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92\") " pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.208976 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.748094 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 16:07:51 crc kubenswrapper[4945]: I1008 16:07:51.797671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92","Type":"ContainerStarted","Data":"f856bc3883a33b018cfefbd4e0aa11545902271c9d0afa39037764722f6f84c8"} Oct 08 16:07:52 crc kubenswrapper[4945]: I1008 16:07:52.351019 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e540f1d4-87f5-4ddf-a9a0-d8141141b9d5" path="/var/lib/kubelet/pods/e540f1d4-87f5-4ddf-a9a0-d8141141b9d5/volumes" Oct 08 16:07:55 crc kubenswrapper[4945]: I1008 16:07:55.024075 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:07:55 crc kubenswrapper[4945]: I1008 16:07:55.840467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"544b7b89e3edd979bb8f670aa366fa3c50d00f52c5c2e2b0b335e2fe3ebd218f"} Oct 08 16:07:55 crc kubenswrapper[4945]: I1008 16:07:55.842970 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92","Type":"ContainerStarted","Data":"d11795e63b959b0d7056eabaa587fb151eb09d0642c84ee2f328105c5939afdb"} Oct 08 16:08:02 crc kubenswrapper[4945]: I1008 16:08:02.913228 4945 generic.go:334] "Generic (PLEG): container finished" podID="9ce8e6f8-1567-4bd4-8f2c-ada70838ec92" containerID="d11795e63b959b0d7056eabaa587fb151eb09d0642c84ee2f328105c5939afdb" exitCode=0 Oct 08 16:08:02 crc kubenswrapper[4945]: I1008 16:08:02.913343 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92","Type":"ContainerDied","Data":"d11795e63b959b0d7056eabaa587fb151eb09d0642c84ee2f328105c5939afdb"} Oct 08 16:08:03 crc kubenswrapper[4945]: I1008 16:08:03.926026 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92","Type":"ContainerStarted","Data":"d97ca9f3fcf653f7ef6154a544ff9b53c0d0884dc2e45e47f30e158e3850820c"} Oct 08 16:08:06 crc kubenswrapper[4945]: I1008 16:08:06.956471 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92","Type":"ContainerStarted","Data":"01f89c0183fcc6c6824710b48d8c0ee20d07fccfe5c71d4f89d4a750c8ebdc70"} Oct 08 16:08:07 crc kubenswrapper[4945]: I1008 16:08:07.966766 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ce8e6f8-1567-4bd4-8f2c-ada70838ec92","Type":"ContainerStarted","Data":"272a164cdb7ef013f3e7aa9e1405c856d5843a02924d6c07f17d8d444546701c"} Oct 08 16:08:07 crc kubenswrapper[4945]: I1008 16:08:07.997740 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.997721955 podStartE2EDuration="17.997721955s" podCreationTimestamp="2025-10-08 16:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 16:08:07.988790025 +0000 UTC m=+3297.342704936" watchObservedRunningTime="2025-10-08 16:08:07.997721955 +0000 UTC m=+3297.351636856" Oct 08 16:08:11 crc kubenswrapper[4945]: I1008 16:08:11.209365 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 16:08:21 crc kubenswrapper[4945]: I1008 16:08:21.210237 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 16:08:21 crc kubenswrapper[4945]: I1008 16:08:21.218407 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 16:08:22 crc kubenswrapper[4945]: I1008 16:08:22.096869 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 16:08:32 crc kubenswrapper[4945]: I1008 16:08:32.867083 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 16:08:32 crc kubenswrapper[4945]: I1008 16:08:32.875475 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 16:08:32 crc kubenswrapper[4945]: I1008 16:08:32.875628 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 16:08:32 crc kubenswrapper[4945]: I1008 16:08:32.879672 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 08 16:08:32 crc kubenswrapper[4945]: I1008 16:08:32.879704 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 08 16:08:32 crc kubenswrapper[4945]: I1008 16:08:32.879878 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 08 16:08:32 crc kubenswrapper[4945]: I1008 16:08:32.880071 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mwxqb" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.058423 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.058672 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xjck\" (UniqueName: \"kubernetes.io/projected/0f3aa285-da71-4df6-8188-cc528008a61b-kube-api-access-8xjck\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.058838 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-config-data\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.058940 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.059068 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.059201 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.059390 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.059540 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.059703 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161054 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161158 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161210 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161251 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161293 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xjck\" (UniqueName: \"kubernetes.io/projected/0f3aa285-da71-4df6-8188-cc528008a61b-kube-api-access-8xjck\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-config-data\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161371 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161408 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161423 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.161629 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.163404 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-config-data\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.163531 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.164158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.166932 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.167618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.191845 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xjck\" (UniqueName: \"kubernetes.io/projected/0f3aa285-da71-4df6-8188-cc528008a61b-kube-api-access-8xjck\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.191869 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.194150 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.222103 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " pod="openstack/tempest-tests-tempest" Oct 08 16:08:33 crc kubenswrapper[4945]: I1008 16:08:33.501758 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 16:08:34 crc kubenswrapper[4945]: I1008 16:08:34.048070 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 16:08:34 crc kubenswrapper[4945]: I1008 16:08:34.206202 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0f3aa285-da71-4df6-8188-cc528008a61b","Type":"ContainerStarted","Data":"59f70af86fe4980a8e7c8c09d07ff378c2e1a936fbd3e6ce91632ac00668e5f2"} Oct 08 16:08:38 crc kubenswrapper[4945]: I1008 16:08:38.679685 4945 scope.go:117] "RemoveContainer" containerID="9334b633912154d5a07675395e572e3e4f7a03ee0b0f11ce71d5126d5f63dcf8" Oct 08 16:08:45 crc kubenswrapper[4945]: I1008 16:08:45.859438 4945 scope.go:117] "RemoveContainer" containerID="f369e2d18bb008538dfcdb3cf115bdbcddab9f8002e86fc74dde30e0774d4bb9" Oct 08 16:08:45 crc kubenswrapper[4945]: I1008 16:08:45.928038 4945 scope.go:117] "RemoveContainer" containerID="25291cef51dbdbe9f63d8e75a049c6415369c172fc97380ccc6035814d3a7927" Oct 08 16:08:49 crc kubenswrapper[4945]: I1008 16:08:49.381398 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0f3aa285-da71-4df6-8188-cc528008a61b","Type":"ContainerStarted","Data":"6b38deae846260c42fa56914bb7126d29bb4c4a9a83e1a25e314279899d5cfb1"} Oct 08 16:08:49 crc kubenswrapper[4945]: I1008 16:08:49.400684 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.60118625 podStartE2EDuration="18.400665512s" podCreationTimestamp="2025-10-08 16:08:31 +0000 UTC" firstStartedPulling="2025-10-08 16:08:34.029566344 +0000 UTC m=+3323.383481245" lastFinishedPulling="2025-10-08 16:08:47.829045606 +0000 UTC m=+3337.182960507" observedRunningTime="2025-10-08 16:08:49.39701385 +0000 UTC m=+3338.750928771" watchObservedRunningTime="2025-10-08 16:08:49.400665512 +0000 UTC m=+3338.754580413" Oct 08 16:10:19 crc kubenswrapper[4945]: I1008 16:10:19.184537 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:10:19 crc kubenswrapper[4945]: I1008 16:10:19.185239 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.519955 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zc449"] Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.525071 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.548555 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zc449"] Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.580458 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-utilities\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.580533 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfxbt\" (UniqueName: \"kubernetes.io/projected/6b37b614-c0b3-4854-b609-01190db91672-kube-api-access-hfxbt\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.580624 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-catalog-content\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.681582 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfxbt\" (UniqueName: \"kubernetes.io/projected/6b37b614-c0b3-4854-b609-01190db91672-kube-api-access-hfxbt\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.681684 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-catalog-content\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.681788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-utilities\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.682256 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-catalog-content\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.682273 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-utilities\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.701472 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfxbt\" (UniqueName: \"kubernetes.io/projected/6b37b614-c0b3-4854-b609-01190db91672-kube-api-access-hfxbt\") pod \"community-operators-zc449\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:45 crc kubenswrapper[4945]: I1008 16:10:45.853665 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:46 crc kubenswrapper[4945]: I1008 16:10:46.412156 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zc449"] Oct 08 16:10:46 crc kubenswrapper[4945]: W1008 16:10:46.413901 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b37b614_c0b3_4854_b609_01190db91672.slice/crio-4877024755c9396596b9b24cc0a4aa21890dc8b9370138f3b00cbefe15ec6987 WatchSource:0}: Error finding container 4877024755c9396596b9b24cc0a4aa21890dc8b9370138f3b00cbefe15ec6987: Status 404 returned error can't find the container with id 4877024755c9396596b9b24cc0a4aa21890dc8b9370138f3b00cbefe15ec6987 Oct 08 16:10:46 crc kubenswrapper[4945]: I1008 16:10:46.585505 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zc449" event={"ID":"6b37b614-c0b3-4854-b609-01190db91672","Type":"ContainerStarted","Data":"4877024755c9396596b9b24cc0a4aa21890dc8b9370138f3b00cbefe15ec6987"} Oct 08 16:10:47 crc kubenswrapper[4945]: I1008 16:10:47.599168 4945 generic.go:334] "Generic (PLEG): container finished" podID="6b37b614-c0b3-4854-b609-01190db91672" containerID="e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f" exitCode=0 Oct 08 16:10:47 crc kubenswrapper[4945]: I1008 16:10:47.599299 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zc449" event={"ID":"6b37b614-c0b3-4854-b609-01190db91672","Type":"ContainerDied","Data":"e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f"} Oct 08 16:10:49 crc kubenswrapper[4945]: I1008 16:10:49.184857 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:10:49 crc kubenswrapper[4945]: I1008 16:10:49.185414 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:10:49 crc kubenswrapper[4945]: I1008 16:10:49.620186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zc449" event={"ID":"6b37b614-c0b3-4854-b609-01190db91672","Type":"ContainerStarted","Data":"b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e"} Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.312577 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nk4f9"] Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.315876 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.322725 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-utilities\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.322862 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-catalog-content\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.322899 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46dxm\" (UniqueName: \"kubernetes.io/projected/16c9460b-517f-400c-a1ee-a046c794790d-kube-api-access-46dxm\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.324239 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk4f9"] Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.425699 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-utilities\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.425832 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-catalog-content\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.425866 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46dxm\" (UniqueName: \"kubernetes.io/projected/16c9460b-517f-400c-a1ee-a046c794790d-kube-api-access-46dxm\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.426296 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-utilities\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.426349 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-catalog-content\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.444987 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46dxm\" (UniqueName: \"kubernetes.io/projected/16c9460b-517f-400c-a1ee-a046c794790d-kube-api-access-46dxm\") pod \"certified-operators-nk4f9\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.640011 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.662722 4945 generic.go:334] "Generic (PLEG): container finished" podID="6b37b614-c0b3-4854-b609-01190db91672" containerID="b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e" exitCode=0 Oct 08 16:10:52 crc kubenswrapper[4945]: I1008 16:10:52.662816 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zc449" event={"ID":"6b37b614-c0b3-4854-b609-01190db91672","Type":"ContainerDied","Data":"b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e"} Oct 08 16:10:53 crc kubenswrapper[4945]: I1008 16:10:53.206540 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk4f9"] Oct 08 16:10:53 crc kubenswrapper[4945]: I1008 16:10:53.676978 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4f9" event={"ID":"16c9460b-517f-400c-a1ee-a046c794790d","Type":"ContainerStarted","Data":"f6604c64426d6aae6df41391dda9a1a9230c53344846665750b4812fc0a51ceb"} Oct 08 16:10:54 crc kubenswrapper[4945]: I1008 16:10:54.688828 4945 generic.go:334] "Generic (PLEG): container finished" podID="16c9460b-517f-400c-a1ee-a046c794790d" containerID="a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c" exitCode=0 Oct 08 16:10:54 crc kubenswrapper[4945]: I1008 16:10:54.688902 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4f9" event={"ID":"16c9460b-517f-400c-a1ee-a046c794790d","Type":"ContainerDied","Data":"a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c"} Oct 08 16:10:54 crc kubenswrapper[4945]: I1008 16:10:54.694685 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zc449" event={"ID":"6b37b614-c0b3-4854-b609-01190db91672","Type":"ContainerStarted","Data":"95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c"} Oct 08 16:10:54 crc kubenswrapper[4945]: I1008 16:10:54.729623 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zc449" podStartSLOduration=3.910827663 podStartE2EDuration="9.729601261s" podCreationTimestamp="2025-10-08 16:10:45 +0000 UTC" firstStartedPulling="2025-10-08 16:10:47.601766364 +0000 UTC m=+3456.955681295" lastFinishedPulling="2025-10-08 16:10:53.420539982 +0000 UTC m=+3462.774454893" observedRunningTime="2025-10-08 16:10:54.719267562 +0000 UTC m=+3464.073182483" watchObservedRunningTime="2025-10-08 16:10:54.729601261 +0000 UTC m=+3464.083516172" Oct 08 16:10:55 crc kubenswrapper[4945]: I1008 16:10:55.854902 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:55 crc kubenswrapper[4945]: I1008 16:10:55.855265 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:55 crc kubenswrapper[4945]: I1008 16:10:55.917953 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:10:56 crc kubenswrapper[4945]: I1008 16:10:56.718462 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4f9" event={"ID":"16c9460b-517f-400c-a1ee-a046c794790d","Type":"ContainerStarted","Data":"be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147"} Oct 08 16:11:02 crc kubenswrapper[4945]: I1008 16:11:02.781264 4945 generic.go:334] "Generic (PLEG): container finished" podID="16c9460b-517f-400c-a1ee-a046c794790d" containerID="be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147" exitCode=0 Oct 08 16:11:02 crc kubenswrapper[4945]: I1008 16:11:02.781354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4f9" event={"ID":"16c9460b-517f-400c-a1ee-a046c794790d","Type":"ContainerDied","Data":"be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147"} Oct 08 16:11:04 crc kubenswrapper[4945]: I1008 16:11:04.802843 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4f9" event={"ID":"16c9460b-517f-400c-a1ee-a046c794790d","Type":"ContainerStarted","Data":"f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275"} Oct 08 16:11:04 crc kubenswrapper[4945]: I1008 16:11:04.824011 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nk4f9" podStartSLOduration=3.898662618 podStartE2EDuration="12.823981045s" podCreationTimestamp="2025-10-08 16:10:52 +0000 UTC" firstStartedPulling="2025-10-08 16:10:54.691646321 +0000 UTC m=+3464.045561222" lastFinishedPulling="2025-10-08 16:11:03.616964708 +0000 UTC m=+3472.970879649" observedRunningTime="2025-10-08 16:11:04.819298909 +0000 UTC m=+3474.173213820" watchObservedRunningTime="2025-10-08 16:11:04.823981045 +0000 UTC m=+3474.177895976" Oct 08 16:11:05 crc kubenswrapper[4945]: I1008 16:11:05.906336 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:11:05 crc kubenswrapper[4945]: I1008 16:11:05.947806 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zc449"] Oct 08 16:11:06 crc kubenswrapper[4945]: I1008 16:11:06.824495 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zc449" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="registry-server" containerID="cri-o://95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c" gracePeriod=2 Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.337449 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.379949 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-catalog-content\") pod \"6b37b614-c0b3-4854-b609-01190db91672\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.380093 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-utilities\") pod \"6b37b614-c0b3-4854-b609-01190db91672\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.380196 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfxbt\" (UniqueName: \"kubernetes.io/projected/6b37b614-c0b3-4854-b609-01190db91672-kube-api-access-hfxbt\") pod \"6b37b614-c0b3-4854-b609-01190db91672\" (UID: \"6b37b614-c0b3-4854-b609-01190db91672\") " Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.381032 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-utilities" (OuterVolumeSpecName: "utilities") pod "6b37b614-c0b3-4854-b609-01190db91672" (UID: "6b37b614-c0b3-4854-b609-01190db91672"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.389712 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b37b614-c0b3-4854-b609-01190db91672-kube-api-access-hfxbt" (OuterVolumeSpecName: "kube-api-access-hfxbt") pod "6b37b614-c0b3-4854-b609-01190db91672" (UID: "6b37b614-c0b3-4854-b609-01190db91672"). InnerVolumeSpecName "kube-api-access-hfxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.426288 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b37b614-c0b3-4854-b609-01190db91672" (UID: "6b37b614-c0b3-4854-b609-01190db91672"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.482506 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.482556 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfxbt\" (UniqueName: \"kubernetes.io/projected/6b37b614-c0b3-4854-b609-01190db91672-kube-api-access-hfxbt\") on node \"crc\" DevicePath \"\"" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.482570 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b37b614-c0b3-4854-b609-01190db91672-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.844803 4945 generic.go:334] "Generic (PLEG): container finished" podID="6b37b614-c0b3-4854-b609-01190db91672" containerID="95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c" exitCode=0 Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.844861 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zc449" event={"ID":"6b37b614-c0b3-4854-b609-01190db91672","Type":"ContainerDied","Data":"95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c"} Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.844893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zc449" event={"ID":"6b37b614-c0b3-4854-b609-01190db91672","Type":"ContainerDied","Data":"4877024755c9396596b9b24cc0a4aa21890dc8b9370138f3b00cbefe15ec6987"} Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.844917 4945 scope.go:117] "RemoveContainer" containerID="95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.844999 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zc449" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.872790 4945 scope.go:117] "RemoveContainer" containerID="b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.878467 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zc449"] Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.886850 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zc449"] Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.910650 4945 scope.go:117] "RemoveContainer" containerID="e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.939208 4945 scope.go:117] "RemoveContainer" containerID="95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c" Oct 08 16:11:07 crc kubenswrapper[4945]: E1008 16:11:07.939624 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c\": container with ID starting with 95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c not found: ID does not exist" containerID="95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.939656 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c"} err="failed to get container status \"95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c\": rpc error: code = NotFound desc = could not find container \"95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c\": container with ID starting with 95fe7f2dd3304a2f71c9536b3285f4ecbf204ee1cf2a373ff4bc2405dc27f81c not found: ID does not exist" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.939681 4945 scope.go:117] "RemoveContainer" containerID="b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e" Oct 08 16:11:07 crc kubenswrapper[4945]: E1008 16:11:07.939913 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e\": container with ID starting with b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e not found: ID does not exist" containerID="b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.939939 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e"} err="failed to get container status \"b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e\": rpc error: code = NotFound desc = could not find container \"b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e\": container with ID starting with b66bb673390c223a500f8241506a9fa207a749f5c41edf79bb976232a5cc2b3e not found: ID does not exist" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.939954 4945 scope.go:117] "RemoveContainer" containerID="e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f" Oct 08 16:11:07 crc kubenswrapper[4945]: E1008 16:11:07.940175 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f\": container with ID starting with e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f not found: ID does not exist" containerID="e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f" Oct 08 16:11:07 crc kubenswrapper[4945]: I1008 16:11:07.940215 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f"} err="failed to get container status \"e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f\": rpc error: code = NotFound desc = could not find container \"e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f\": container with ID starting with e5c403f4280e0ea04066f13c49845866c0877ae35edda92029ac5b5cb8c4b84f not found: ID does not exist" Oct 08 16:11:08 crc kubenswrapper[4945]: I1008 16:11:08.035111 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b37b614-c0b3-4854-b609-01190db91672" path="/var/lib/kubelet/pods/6b37b614-c0b3-4854-b609-01190db91672/volumes" Oct 08 16:11:12 crc kubenswrapper[4945]: I1008 16:11:12.641253 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:11:12 crc kubenswrapper[4945]: I1008 16:11:12.641794 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:11:12 crc kubenswrapper[4945]: I1008 16:11:12.687223 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:11:12 crc kubenswrapper[4945]: I1008 16:11:12.939194 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:11:12 crc kubenswrapper[4945]: I1008 16:11:12.997637 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nk4f9"] Oct 08 16:11:14 crc kubenswrapper[4945]: I1008 16:11:14.914906 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nk4f9" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="registry-server" containerID="cri-o://f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275" gracePeriod=2 Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.648607 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.839664 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-catalog-content\") pod \"16c9460b-517f-400c-a1ee-a046c794790d\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.839826 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-utilities\") pod \"16c9460b-517f-400c-a1ee-a046c794790d\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.840709 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-utilities" (OuterVolumeSpecName: "utilities") pod "16c9460b-517f-400c-a1ee-a046c794790d" (UID: "16c9460b-517f-400c-a1ee-a046c794790d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.841151 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46dxm\" (UniqueName: \"kubernetes.io/projected/16c9460b-517f-400c-a1ee-a046c794790d-kube-api-access-46dxm\") pod \"16c9460b-517f-400c-a1ee-a046c794790d\" (UID: \"16c9460b-517f-400c-a1ee-a046c794790d\") " Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.842715 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.846603 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c9460b-517f-400c-a1ee-a046c794790d-kube-api-access-46dxm" (OuterVolumeSpecName: "kube-api-access-46dxm") pod "16c9460b-517f-400c-a1ee-a046c794790d" (UID: "16c9460b-517f-400c-a1ee-a046c794790d"). InnerVolumeSpecName "kube-api-access-46dxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.900855 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16c9460b-517f-400c-a1ee-a046c794790d" (UID: "16c9460b-517f-400c-a1ee-a046c794790d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.926056 4945 generic.go:334] "Generic (PLEG): container finished" podID="16c9460b-517f-400c-a1ee-a046c794790d" containerID="f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275" exitCode=0 Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.926103 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4f9" event={"ID":"16c9460b-517f-400c-a1ee-a046c794790d","Type":"ContainerDied","Data":"f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275"} Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.926158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4f9" event={"ID":"16c9460b-517f-400c-a1ee-a046c794790d","Type":"ContainerDied","Data":"f6604c64426d6aae6df41391dda9a1a9230c53344846665750b4812fc0a51ceb"} Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.926155 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk4f9" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.926174 4945 scope.go:117] "RemoveContainer" containerID="f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.943933 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46dxm\" (UniqueName: \"kubernetes.io/projected/16c9460b-517f-400c-a1ee-a046c794790d-kube-api-access-46dxm\") on node \"crc\" DevicePath \"\"" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.943966 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16c9460b-517f-400c-a1ee-a046c794790d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.959515 4945 scope.go:117] "RemoveContainer" containerID="be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147" Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.964708 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nk4f9"] Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.975096 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nk4f9"] Oct 08 16:11:15 crc kubenswrapper[4945]: I1008 16:11:15.980299 4945 scope.go:117] "RemoveContainer" containerID="a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c" Oct 08 16:11:16 crc kubenswrapper[4945]: I1008 16:11:16.029426 4945 scope.go:117] "RemoveContainer" containerID="f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275" Oct 08 16:11:16 crc kubenswrapper[4945]: E1008 16:11:16.030385 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275\": container with ID starting with f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275 not found: ID does not exist" containerID="f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275" Oct 08 16:11:16 crc kubenswrapper[4945]: I1008 16:11:16.030569 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275"} err="failed to get container status \"f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275\": rpc error: code = NotFound desc = could not find container \"f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275\": container with ID starting with f1518afe21aa03136703b229568bc86991e8bb04ac2a8d249320dbad93a40275 not found: ID does not exist" Oct 08 16:11:16 crc kubenswrapper[4945]: I1008 16:11:16.030676 4945 scope.go:117] "RemoveContainer" containerID="be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147" Oct 08 16:11:16 crc kubenswrapper[4945]: E1008 16:11:16.031290 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147\": container with ID starting with be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147 not found: ID does not exist" containerID="be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147" Oct 08 16:11:16 crc kubenswrapper[4945]: I1008 16:11:16.031330 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147"} err="failed to get container status \"be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147\": rpc error: code = NotFound desc = could not find container \"be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147\": container with ID starting with be64084ea3703d8c70e96316830201ba98b2ba76b641c36e49a2144e4ab44147 not found: ID does not exist" Oct 08 16:11:16 crc kubenswrapper[4945]: I1008 16:11:16.031357 4945 scope.go:117] "RemoveContainer" containerID="a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c" Oct 08 16:11:16 crc kubenswrapper[4945]: E1008 16:11:16.031704 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c\": container with ID starting with a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c not found: ID does not exist" containerID="a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c" Oct 08 16:11:16 crc kubenswrapper[4945]: I1008 16:11:16.031824 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c"} err="failed to get container status \"a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c\": rpc error: code = NotFound desc = could not find container \"a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c\": container with ID starting with a22af4ad2acd7cbbb57bfbe7cdbb52e11177f5afa58731a4260e6204f7a4fc6c not found: ID does not exist" Oct 08 16:11:16 crc kubenswrapper[4945]: I1008 16:11:16.036886 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c9460b-517f-400c-a1ee-a046c794790d" path="/var/lib/kubelet/pods/16c9460b-517f-400c-a1ee-a046c794790d/volumes" Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.184343 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.184743 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.184791 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.185625 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"544b7b89e3edd979bb8f670aa366fa3c50d00f52c5c2e2b0b335e2fe3ebd218f"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.185694 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://544b7b89e3edd979bb8f670aa366fa3c50d00f52c5c2e2b0b335e2fe3ebd218f" gracePeriod=600 Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.973521 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="544b7b89e3edd979bb8f670aa366fa3c50d00f52c5c2e2b0b335e2fe3ebd218f" exitCode=0 Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.973581 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"544b7b89e3edd979bb8f670aa366fa3c50d00f52c5c2e2b0b335e2fe3ebd218f"} Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.973891 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439"} Oct 08 16:11:19 crc kubenswrapper[4945]: I1008 16:11:19.973916 4945 scope.go:117] "RemoveContainer" containerID="38e63455c90d4de7e22ea1883c2a902de2c6ea33de0c2909d982ce2215cf1609" Oct 08 16:13:19 crc kubenswrapper[4945]: I1008 16:13:19.183948 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:13:19 crc kubenswrapper[4945]: I1008 16:13:19.184595 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.230884 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-98z8c"] Oct 08 16:13:42 crc kubenswrapper[4945]: E1008 16:13:42.252963 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="extract-utilities" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.253008 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="extract-utilities" Oct 08 16:13:42 crc kubenswrapper[4945]: E1008 16:13:42.253038 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="extract-content" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.253047 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="extract-content" Oct 08 16:13:42 crc kubenswrapper[4945]: E1008 16:13:42.253091 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="registry-server" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.253100 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="registry-server" Oct 08 16:13:42 crc kubenswrapper[4945]: E1008 16:13:42.253158 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="extract-content" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.253169 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="extract-content" Oct 08 16:13:42 crc kubenswrapper[4945]: E1008 16:13:42.253207 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="registry-server" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.253217 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="registry-server" Oct 08 16:13:42 crc kubenswrapper[4945]: E1008 16:13:42.253250 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="extract-utilities" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.253259 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="extract-utilities" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.254153 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c9460b-517f-400c-a1ee-a046c794790d" containerName="registry-server" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.254208 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b37b614-c0b3-4854-b609-01190db91672" containerName="registry-server" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.269067 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.291772 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-98z8c"] Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.342317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqzx2\" (UniqueName: \"kubernetes.io/projected/39850f35-424d-409d-9400-4714252a7f06-kube-api-access-vqzx2\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.342619 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-catalog-content\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.342680 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-utilities\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.445419 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-catalog-content\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.445478 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-utilities\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.445605 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqzx2\" (UniqueName: \"kubernetes.io/projected/39850f35-424d-409d-9400-4714252a7f06-kube-api-access-vqzx2\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.445978 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-catalog-content\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.446376 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-utilities\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.465092 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqzx2\" (UniqueName: \"kubernetes.io/projected/39850f35-424d-409d-9400-4714252a7f06-kube-api-access-vqzx2\") pod \"redhat-operators-98z8c\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:42 crc kubenswrapper[4945]: I1008 16:13:42.597828 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:13:43 crc kubenswrapper[4945]: I1008 16:13:43.092503 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-98z8c"] Oct 08 16:13:43 crc kubenswrapper[4945]: I1008 16:13:43.331234 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-98z8c" event={"ID":"39850f35-424d-409d-9400-4714252a7f06","Type":"ContainerStarted","Data":"5ed08e646dd3dc2b7efb5a25e869328e8271aaae31de3039169882e80dcab4ee"} Oct 08 16:13:44 crc kubenswrapper[4945]: I1008 16:13:44.356379 4945 generic.go:334] "Generic (PLEG): container finished" podID="39850f35-424d-409d-9400-4714252a7f06" containerID="8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71" exitCode=0 Oct 08 16:13:44 crc kubenswrapper[4945]: I1008 16:13:44.356420 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-98z8c" event={"ID":"39850f35-424d-409d-9400-4714252a7f06","Type":"ContainerDied","Data":"8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71"} Oct 08 16:13:44 crc kubenswrapper[4945]: I1008 16:13:44.360390 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 16:13:46 crc kubenswrapper[4945]: I1008 16:13:46.377961 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-98z8c" event={"ID":"39850f35-424d-409d-9400-4714252a7f06","Type":"ContainerStarted","Data":"0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773"} Oct 08 16:13:49 crc kubenswrapper[4945]: I1008 16:13:49.184427 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:13:49 crc kubenswrapper[4945]: I1008 16:13:49.184970 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:13:51 crc kubenswrapper[4945]: I1008 16:13:51.429483 4945 generic.go:334] "Generic (PLEG): container finished" podID="39850f35-424d-409d-9400-4714252a7f06" containerID="0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773" exitCode=0 Oct 08 16:13:51 crc kubenswrapper[4945]: I1008 16:13:51.429553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-98z8c" event={"ID":"39850f35-424d-409d-9400-4714252a7f06","Type":"ContainerDied","Data":"0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773"} Oct 08 16:13:53 crc kubenswrapper[4945]: I1008 16:13:53.452837 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-98z8c" event={"ID":"39850f35-424d-409d-9400-4714252a7f06","Type":"ContainerStarted","Data":"af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489"} Oct 08 16:13:53 crc kubenswrapper[4945]: I1008 16:13:53.475998 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-98z8c" podStartSLOduration=2.885660135 podStartE2EDuration="11.475981537s" podCreationTimestamp="2025-10-08 16:13:42 +0000 UTC" firstStartedPulling="2025-10-08 16:13:44.360093003 +0000 UTC m=+3633.714007904" lastFinishedPulling="2025-10-08 16:13:52.950414405 +0000 UTC m=+3642.304329306" observedRunningTime="2025-10-08 16:13:53.471546186 +0000 UTC m=+3642.825461127" watchObservedRunningTime="2025-10-08 16:13:53.475981537 +0000 UTC m=+3642.829896438" Oct 08 16:14:02 crc kubenswrapper[4945]: I1008 16:14:02.598572 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:14:02 crc kubenswrapper[4945]: I1008 16:14:02.599269 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:14:03 crc kubenswrapper[4945]: I1008 16:14:03.649563 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-98z8c" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="registry-server" probeResult="failure" output=< Oct 08 16:14:03 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:14:03 crc kubenswrapper[4945]: > Oct 08 16:14:12 crc kubenswrapper[4945]: I1008 16:14:12.646796 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:14:12 crc kubenswrapper[4945]: I1008 16:14:12.694912 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:14:13 crc kubenswrapper[4945]: I1008 16:14:13.433109 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-98z8c"] Oct 08 16:14:14 crc kubenswrapper[4945]: I1008 16:14:14.669424 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-98z8c" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="registry-server" containerID="cri-o://af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489" gracePeriod=2 Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.145382 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.233815 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-catalog-content\") pod \"39850f35-424d-409d-9400-4714252a7f06\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.234018 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqzx2\" (UniqueName: \"kubernetes.io/projected/39850f35-424d-409d-9400-4714252a7f06-kube-api-access-vqzx2\") pod \"39850f35-424d-409d-9400-4714252a7f06\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.234045 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-utilities\") pod \"39850f35-424d-409d-9400-4714252a7f06\" (UID: \"39850f35-424d-409d-9400-4714252a7f06\") " Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.234833 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-utilities" (OuterVolumeSpecName: "utilities") pod "39850f35-424d-409d-9400-4714252a7f06" (UID: "39850f35-424d-409d-9400-4714252a7f06"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.240556 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39850f35-424d-409d-9400-4714252a7f06-kube-api-access-vqzx2" (OuterVolumeSpecName: "kube-api-access-vqzx2") pod "39850f35-424d-409d-9400-4714252a7f06" (UID: "39850f35-424d-409d-9400-4714252a7f06"). InnerVolumeSpecName "kube-api-access-vqzx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.326288 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39850f35-424d-409d-9400-4714252a7f06" (UID: "39850f35-424d-409d-9400-4714252a7f06"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.335959 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.335986 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqzx2\" (UniqueName: \"kubernetes.io/projected/39850f35-424d-409d-9400-4714252a7f06-kube-api-access-vqzx2\") on node \"crc\" DevicePath \"\"" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.335997 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39850f35-424d-409d-9400-4714252a7f06-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.680641 4945 generic.go:334] "Generic (PLEG): container finished" podID="39850f35-424d-409d-9400-4714252a7f06" containerID="af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489" exitCode=0 Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.680697 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-98z8c" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.680716 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-98z8c" event={"ID":"39850f35-424d-409d-9400-4714252a7f06","Type":"ContainerDied","Data":"af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489"} Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.681000 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-98z8c" event={"ID":"39850f35-424d-409d-9400-4714252a7f06","Type":"ContainerDied","Data":"5ed08e646dd3dc2b7efb5a25e869328e8271aaae31de3039169882e80dcab4ee"} Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.681020 4945 scope.go:117] "RemoveContainer" containerID="af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.714582 4945 scope.go:117] "RemoveContainer" containerID="0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.725687 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-98z8c"] Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.736333 4945 scope.go:117] "RemoveContainer" containerID="8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.737912 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-98z8c"] Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.778358 4945 scope.go:117] "RemoveContainer" containerID="af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489" Oct 08 16:14:15 crc kubenswrapper[4945]: E1008 16:14:15.778754 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489\": container with ID starting with af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489 not found: ID does not exist" containerID="af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.778795 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489"} err="failed to get container status \"af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489\": rpc error: code = NotFound desc = could not find container \"af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489\": container with ID starting with af2288c37ef98331bfe9849ea36afeb4a7d980202ca056ab09aa1e0459ad9489 not found: ID does not exist" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.778823 4945 scope.go:117] "RemoveContainer" containerID="0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773" Oct 08 16:14:15 crc kubenswrapper[4945]: E1008 16:14:15.779164 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773\": container with ID starting with 0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773 not found: ID does not exist" containerID="0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.779219 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773"} err="failed to get container status \"0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773\": rpc error: code = NotFound desc = could not find container \"0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773\": container with ID starting with 0d300ec7f208ef3c029864233a205218d584c4967c1c219b5a64f04141cef773 not found: ID does not exist" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.779251 4945 scope.go:117] "RemoveContainer" containerID="8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71" Oct 08 16:14:15 crc kubenswrapper[4945]: E1008 16:14:15.779490 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71\": container with ID starting with 8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71 not found: ID does not exist" containerID="8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71" Oct 08 16:14:15 crc kubenswrapper[4945]: I1008 16:14:15.779510 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71"} err="failed to get container status \"8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71\": rpc error: code = NotFound desc = could not find container \"8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71\": container with ID starting with 8ca6496e5a28ba2bbd72cee83d1b5c9448cba84ec1ce51b331dca35d0104fd71 not found: ID does not exist" Oct 08 16:14:16 crc kubenswrapper[4945]: I1008 16:14:16.036375 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39850f35-424d-409d-9400-4714252a7f06" path="/var/lib/kubelet/pods/39850f35-424d-409d-9400-4714252a7f06/volumes" Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.184672 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.185199 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.185246 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.186037 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.186087 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" gracePeriod=600 Oct 08 16:14:19 crc kubenswrapper[4945]: E1008 16:14:19.306723 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.724506 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" exitCode=0 Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.724555 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439"} Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.724592 4945 scope.go:117] "RemoveContainer" containerID="544b7b89e3edd979bb8f670aa366fa3c50d00f52c5c2e2b0b335e2fe3ebd218f" Oct 08 16:14:19 crc kubenswrapper[4945]: I1008 16:14:19.725404 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:14:19 crc kubenswrapper[4945]: E1008 16:14:19.725738 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:14:32 crc kubenswrapper[4945]: I1008 16:14:32.030829 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:14:32 crc kubenswrapper[4945]: E1008 16:14:32.031811 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:14:47 crc kubenswrapper[4945]: I1008 16:14:47.024483 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:14:47 crc kubenswrapper[4945]: E1008 16:14:47.025341 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:14:58 crc kubenswrapper[4945]: I1008 16:14:58.023953 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:14:58 crc kubenswrapper[4945]: E1008 16:14:58.024968 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.154862 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm"] Oct 08 16:15:00 crc kubenswrapper[4945]: E1008 16:15:00.155325 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="registry-server" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.155338 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="registry-server" Oct 08 16:15:00 crc kubenswrapper[4945]: E1008 16:15:00.155358 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="extract-utilities" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.155365 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="extract-utilities" Oct 08 16:15:00 crc kubenswrapper[4945]: E1008 16:15:00.155411 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="extract-content" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.155422 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="extract-content" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.155651 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="39850f35-424d-409d-9400-4714252a7f06" containerName="registry-server" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.156324 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.158597 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.158629 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.169075 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm"] Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.261287 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7f09c79-5895-41ac-830e-eb717471c62b-config-volume\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.261938 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7f09c79-5895-41ac-830e-eb717471c62b-secret-volume\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.262236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9mtd\" (UniqueName: \"kubernetes.io/projected/e7f09c79-5895-41ac-830e-eb717471c62b-kube-api-access-g9mtd\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.364785 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7f09c79-5895-41ac-830e-eb717471c62b-config-volume\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.364847 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7f09c79-5895-41ac-830e-eb717471c62b-secret-volume\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.364919 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9mtd\" (UniqueName: \"kubernetes.io/projected/e7f09c79-5895-41ac-830e-eb717471c62b-kube-api-access-g9mtd\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.366168 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7f09c79-5895-41ac-830e-eb717471c62b-config-volume\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.374490 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7f09c79-5895-41ac-830e-eb717471c62b-secret-volume\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.381662 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9mtd\" (UniqueName: \"kubernetes.io/projected/e7f09c79-5895-41ac-830e-eb717471c62b-kube-api-access-g9mtd\") pod \"collect-profiles-29332335-whpgm\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.480705 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:00 crc kubenswrapper[4945]: I1008 16:15:00.950037 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm"] Oct 08 16:15:01 crc kubenswrapper[4945]: I1008 16:15:01.091706 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" event={"ID":"e7f09c79-5895-41ac-830e-eb717471c62b","Type":"ContainerStarted","Data":"90e164e2524d2c55809a0c35f39f92d7a090d9d5049ebab78a8194b0e20594b1"} Oct 08 16:15:02 crc kubenswrapper[4945]: I1008 16:15:02.100884 4945 generic.go:334] "Generic (PLEG): container finished" podID="e7f09c79-5895-41ac-830e-eb717471c62b" containerID="832a4c74370750c767b7575d8dbf1ce38f0bb75cc2317ae337815f30e9582d3b" exitCode=0 Oct 08 16:15:02 crc kubenswrapper[4945]: I1008 16:15:02.101007 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" event={"ID":"e7f09c79-5895-41ac-830e-eb717471c62b","Type":"ContainerDied","Data":"832a4c74370750c767b7575d8dbf1ce38f0bb75cc2317ae337815f30e9582d3b"} Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.541512 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.638936 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9mtd\" (UniqueName: \"kubernetes.io/projected/e7f09c79-5895-41ac-830e-eb717471c62b-kube-api-access-g9mtd\") pod \"e7f09c79-5895-41ac-830e-eb717471c62b\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.638990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7f09c79-5895-41ac-830e-eb717471c62b-config-volume\") pod \"e7f09c79-5895-41ac-830e-eb717471c62b\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.639293 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7f09c79-5895-41ac-830e-eb717471c62b-secret-volume\") pod \"e7f09c79-5895-41ac-830e-eb717471c62b\" (UID: \"e7f09c79-5895-41ac-830e-eb717471c62b\") " Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.639857 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f09c79-5895-41ac-830e-eb717471c62b-config-volume" (OuterVolumeSpecName: "config-volume") pod "e7f09c79-5895-41ac-830e-eb717471c62b" (UID: "e7f09c79-5895-41ac-830e-eb717471c62b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.644914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f09c79-5895-41ac-830e-eb717471c62b-kube-api-access-g9mtd" (OuterVolumeSpecName: "kube-api-access-g9mtd") pod "e7f09c79-5895-41ac-830e-eb717471c62b" (UID: "e7f09c79-5895-41ac-830e-eb717471c62b"). InnerVolumeSpecName "kube-api-access-g9mtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.645038 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f09c79-5895-41ac-830e-eb717471c62b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e7f09c79-5895-41ac-830e-eb717471c62b" (UID: "e7f09c79-5895-41ac-830e-eb717471c62b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.741750 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7f09c79-5895-41ac-830e-eb717471c62b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.741792 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9mtd\" (UniqueName: \"kubernetes.io/projected/e7f09c79-5895-41ac-830e-eb717471c62b-kube-api-access-g9mtd\") on node \"crc\" DevicePath \"\"" Oct 08 16:15:03 crc kubenswrapper[4945]: I1008 16:15:03.741801 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7f09c79-5895-41ac-830e-eb717471c62b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:15:04 crc kubenswrapper[4945]: I1008 16:15:04.138040 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" event={"ID":"e7f09c79-5895-41ac-830e-eb717471c62b","Type":"ContainerDied","Data":"90e164e2524d2c55809a0c35f39f92d7a090d9d5049ebab78a8194b0e20594b1"} Oct 08 16:15:04 crc kubenswrapper[4945]: I1008 16:15:04.138440 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90e164e2524d2c55809a0c35f39f92d7a090d9d5049ebab78a8194b0e20594b1" Oct 08 16:15:04 crc kubenswrapper[4945]: I1008 16:15:04.138095 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm" Oct 08 16:15:04 crc kubenswrapper[4945]: I1008 16:15:04.619608 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t"] Oct 08 16:15:04 crc kubenswrapper[4945]: I1008 16:15:04.628596 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-hcp5t"] Oct 08 16:15:06 crc kubenswrapper[4945]: I1008 16:15:06.052589 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a1c1792-e047-413d-8667-791d14563fb1" path="/var/lib/kubelet/pods/1a1c1792-e047-413d-8667-791d14563fb1/volumes" Oct 08 16:15:13 crc kubenswrapper[4945]: I1008 16:15:13.024245 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:15:13 crc kubenswrapper[4945]: E1008 16:15:13.025442 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:15:25 crc kubenswrapper[4945]: I1008 16:15:25.024472 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:15:25 crc kubenswrapper[4945]: E1008 16:15:25.025646 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:15:38 crc kubenswrapper[4945]: I1008 16:15:38.024364 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:15:38 crc kubenswrapper[4945]: E1008 16:15:38.025058 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:15:46 crc kubenswrapper[4945]: I1008 16:15:46.244385 4945 scope.go:117] "RemoveContainer" containerID="e2d71abcab67512055d8b9de1a1e952ed6b42795b5b10dcfb45aeee58c5584e3" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.613840 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xsgm9"] Oct 08 16:15:48 crc kubenswrapper[4945]: E1008 16:15:48.614806 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f09c79-5895-41ac-830e-eb717471c62b" containerName="collect-profiles" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.614821 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f09c79-5895-41ac-830e-eb717471c62b" containerName="collect-profiles" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.615049 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f09c79-5895-41ac-830e-eb717471c62b" containerName="collect-profiles" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.616551 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.626239 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsgm9"] Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.753621 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k4sw\" (UniqueName: \"kubernetes.io/projected/afc50f02-6cbd-462b-9b74-5079b6cb751d-kube-api-access-6k4sw\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.753713 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-utilities\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.753804 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-catalog-content\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.856541 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k4sw\" (UniqueName: \"kubernetes.io/projected/afc50f02-6cbd-462b-9b74-5079b6cb751d-kube-api-access-6k4sw\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.856918 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-utilities\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.857001 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-catalog-content\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.857433 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-utilities\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.857627 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-catalog-content\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.878950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k4sw\" (UniqueName: \"kubernetes.io/projected/afc50f02-6cbd-462b-9b74-5079b6cb751d-kube-api-access-6k4sw\") pod \"redhat-marketplace-xsgm9\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:48 crc kubenswrapper[4945]: I1008 16:15:48.944155 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:49 crc kubenswrapper[4945]: I1008 16:15:49.025175 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:15:49 crc kubenswrapper[4945]: E1008 16:15:49.025478 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:15:49 crc kubenswrapper[4945]: I1008 16:15:49.466269 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsgm9"] Oct 08 16:15:49 crc kubenswrapper[4945]: I1008 16:15:49.565538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsgm9" event={"ID":"afc50f02-6cbd-462b-9b74-5079b6cb751d","Type":"ContainerStarted","Data":"30b63d0190c04380f9f462497bbcf0f800b7d468931f444af2539acdbb45f8c1"} Oct 08 16:15:50 crc kubenswrapper[4945]: I1008 16:15:50.575898 4945 generic.go:334] "Generic (PLEG): container finished" podID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerID="f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07" exitCode=0 Oct 08 16:15:50 crc kubenswrapper[4945]: I1008 16:15:50.575976 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsgm9" event={"ID":"afc50f02-6cbd-462b-9b74-5079b6cb751d","Type":"ContainerDied","Data":"f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07"} Oct 08 16:15:51 crc kubenswrapper[4945]: I1008 16:15:51.589253 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsgm9" event={"ID":"afc50f02-6cbd-462b-9b74-5079b6cb751d","Type":"ContainerStarted","Data":"a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14"} Oct 08 16:15:52 crc kubenswrapper[4945]: I1008 16:15:52.601497 4945 generic.go:334] "Generic (PLEG): container finished" podID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerID="a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14" exitCode=0 Oct 08 16:15:52 crc kubenswrapper[4945]: I1008 16:15:52.601554 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsgm9" event={"ID":"afc50f02-6cbd-462b-9b74-5079b6cb751d","Type":"ContainerDied","Data":"a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14"} Oct 08 16:15:53 crc kubenswrapper[4945]: I1008 16:15:53.614321 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsgm9" event={"ID":"afc50f02-6cbd-462b-9b74-5079b6cb751d","Type":"ContainerStarted","Data":"a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a"} Oct 08 16:15:53 crc kubenswrapper[4945]: I1008 16:15:53.642993 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xsgm9" podStartSLOduration=3.179291616 podStartE2EDuration="5.642972896s" podCreationTimestamp="2025-10-08 16:15:48 +0000 UTC" firstStartedPulling="2025-10-08 16:15:50.578090042 +0000 UTC m=+3759.932004943" lastFinishedPulling="2025-10-08 16:15:53.041771332 +0000 UTC m=+3762.395686223" observedRunningTime="2025-10-08 16:15:53.634855945 +0000 UTC m=+3762.988770846" watchObservedRunningTime="2025-10-08 16:15:53.642972896 +0000 UTC m=+3762.996887797" Oct 08 16:15:58 crc kubenswrapper[4945]: I1008 16:15:58.944814 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:58 crc kubenswrapper[4945]: I1008 16:15:58.945353 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:58 crc kubenswrapper[4945]: I1008 16:15:58.998435 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:59 crc kubenswrapper[4945]: I1008 16:15:59.713442 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:15:59 crc kubenswrapper[4945]: I1008 16:15:59.759385 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsgm9"] Oct 08 16:16:01 crc kubenswrapper[4945]: I1008 16:16:01.682736 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xsgm9" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="registry-server" containerID="cri-o://a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a" gracePeriod=2 Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.225651 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.356086 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-catalog-content\") pod \"afc50f02-6cbd-462b-9b74-5079b6cb751d\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.356341 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k4sw\" (UniqueName: \"kubernetes.io/projected/afc50f02-6cbd-462b-9b74-5079b6cb751d-kube-api-access-6k4sw\") pod \"afc50f02-6cbd-462b-9b74-5079b6cb751d\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.356478 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-utilities\") pod \"afc50f02-6cbd-462b-9b74-5079b6cb751d\" (UID: \"afc50f02-6cbd-462b-9b74-5079b6cb751d\") " Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.357419 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-utilities" (OuterVolumeSpecName: "utilities") pod "afc50f02-6cbd-462b-9b74-5079b6cb751d" (UID: "afc50f02-6cbd-462b-9b74-5079b6cb751d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.370371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afc50f02-6cbd-462b-9b74-5079b6cb751d" (UID: "afc50f02-6cbd-462b-9b74-5079b6cb751d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.459665 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.459711 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc50f02-6cbd-462b-9b74-5079b6cb751d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.694621 4945 generic.go:334] "Generic (PLEG): container finished" podID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerID="a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a" exitCode=0 Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.694686 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsgm9" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.694691 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsgm9" event={"ID":"afc50f02-6cbd-462b-9b74-5079b6cb751d","Type":"ContainerDied","Data":"a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a"} Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.694765 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsgm9" event={"ID":"afc50f02-6cbd-462b-9b74-5079b6cb751d","Type":"ContainerDied","Data":"30b63d0190c04380f9f462497bbcf0f800b7d468931f444af2539acdbb45f8c1"} Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.694787 4945 scope.go:117] "RemoveContainer" containerID="a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.717727 4945 scope.go:117] "RemoveContainer" containerID="a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.912688 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc50f02-6cbd-462b-9b74-5079b6cb751d-kube-api-access-6k4sw" (OuterVolumeSpecName: "kube-api-access-6k4sw") pod "afc50f02-6cbd-462b-9b74-5079b6cb751d" (UID: "afc50f02-6cbd-462b-9b74-5079b6cb751d"). InnerVolumeSpecName "kube-api-access-6k4sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.932228 4945 scope.go:117] "RemoveContainer" containerID="f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07" Oct 08 16:16:02 crc kubenswrapper[4945]: I1008 16:16:02.971379 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k4sw\" (UniqueName: \"kubernetes.io/projected/afc50f02-6cbd-462b-9b74-5079b6cb751d-kube-api-access-6k4sw\") on node \"crc\" DevicePath \"\"" Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.026960 4945 scope.go:117] "RemoveContainer" containerID="a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a" Oct 08 16:16:03 crc kubenswrapper[4945]: E1008 16:16:03.027485 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a\": container with ID starting with a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a not found: ID does not exist" containerID="a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a" Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.027515 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a"} err="failed to get container status \"a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a\": rpc error: code = NotFound desc = could not find container \"a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a\": container with ID starting with a3f2f23b75532b479347feb6def868f8c226eb207095f7245698f841944e032a not found: ID does not exist" Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.027533 4945 scope.go:117] "RemoveContainer" containerID="a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14" Oct 08 16:16:03 crc kubenswrapper[4945]: E1008 16:16:03.027795 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14\": container with ID starting with a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14 not found: ID does not exist" containerID="a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14" Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.027816 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14"} err="failed to get container status \"a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14\": rpc error: code = NotFound desc = could not find container \"a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14\": container with ID starting with a7e24a7b800b977b2985748f8c6e62ef61b52db1bbd821866ad461a4887d7d14 not found: ID does not exist" Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.027830 4945 scope.go:117] "RemoveContainer" containerID="f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07" Oct 08 16:16:03 crc kubenswrapper[4945]: E1008 16:16:03.028043 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07\": container with ID starting with f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07 not found: ID does not exist" containerID="f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07" Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.028068 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07"} err="failed to get container status \"f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07\": rpc error: code = NotFound desc = could not find container \"f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07\": container with ID starting with f30b8997ddf03a0299a8a2eb9b861de357846591f81f8a414ac974b09aa3bb07 not found: ID does not exist" Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.078208 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsgm9"] Oct 08 16:16:03 crc kubenswrapper[4945]: I1008 16:16:03.087345 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsgm9"] Oct 08 16:16:04 crc kubenswrapper[4945]: I1008 16:16:04.024562 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:16:04 crc kubenswrapper[4945]: E1008 16:16:04.024875 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:16:04 crc kubenswrapper[4945]: I1008 16:16:04.036155 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" path="/var/lib/kubelet/pods/afc50f02-6cbd-462b-9b74-5079b6cb751d/volumes" Oct 08 16:16:17 crc kubenswrapper[4945]: I1008 16:16:17.027590 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:16:17 crc kubenswrapper[4945]: E1008 16:16:17.028403 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:16:31 crc kubenswrapper[4945]: I1008 16:16:31.025019 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:16:31 crc kubenswrapper[4945]: E1008 16:16:31.026438 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:16:43 crc kubenswrapper[4945]: I1008 16:16:43.024936 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:16:43 crc kubenswrapper[4945]: E1008 16:16:43.026334 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:16:58 crc kubenswrapper[4945]: I1008 16:16:58.024229 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:16:58 crc kubenswrapper[4945]: E1008 16:16:58.025201 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:17:11 crc kubenswrapper[4945]: I1008 16:17:11.024636 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:17:11 crc kubenswrapper[4945]: E1008 16:17:11.025445 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:17:26 crc kubenswrapper[4945]: I1008 16:17:26.024541 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:17:26 crc kubenswrapper[4945]: E1008 16:17:26.026908 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:17:39 crc kubenswrapper[4945]: I1008 16:17:39.024663 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:17:39 crc kubenswrapper[4945]: E1008 16:17:39.025552 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:17:50 crc kubenswrapper[4945]: I1008 16:17:50.025667 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:17:50 crc kubenswrapper[4945]: E1008 16:17:50.027496 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:18:01 crc kubenswrapper[4945]: I1008 16:18:01.024440 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:18:01 crc kubenswrapper[4945]: E1008 16:18:01.025247 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:18:16 crc kubenswrapper[4945]: I1008 16:18:16.025816 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:18:16 crc kubenswrapper[4945]: E1008 16:18:16.026615 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:18:27 crc kubenswrapper[4945]: I1008 16:18:27.024835 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:18:27 crc kubenswrapper[4945]: E1008 16:18:27.026323 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:18:40 crc kubenswrapper[4945]: I1008 16:18:40.023953 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:18:40 crc kubenswrapper[4945]: E1008 16:18:40.024637 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:18:55 crc kubenswrapper[4945]: I1008 16:18:55.024380 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:18:55 crc kubenswrapper[4945]: E1008 16:18:55.025395 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:19:08 crc kubenswrapper[4945]: I1008 16:19:08.024786 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:19:08 crc kubenswrapper[4945]: E1008 16:19:08.025552 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:19:23 crc kubenswrapper[4945]: I1008 16:19:23.024165 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:19:23 crc kubenswrapper[4945]: I1008 16:19:23.623575 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"4f31cb254505165039065bc9149b5d3a99d5daf79526310144bd8b0a16330be4"} Oct 08 16:21:49 crc kubenswrapper[4945]: I1008 16:21:49.183996 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:21:49 crc kubenswrapper[4945]: I1008 16:21:49.185090 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:22:19 crc kubenswrapper[4945]: I1008 16:22:19.184660 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:22:19 crc kubenswrapper[4945]: I1008 16:22:19.185238 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.184657 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.185308 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.185363 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.186146 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4f31cb254505165039065bc9149b5d3a99d5daf79526310144bd8b0a16330be4"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.186202 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://4f31cb254505165039065bc9149b5d3a99d5daf79526310144bd8b0a16330be4" gracePeriod=600 Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.597689 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="4f31cb254505165039065bc9149b5d3a99d5daf79526310144bd8b0a16330be4" exitCode=0 Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.597772 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"4f31cb254505165039065bc9149b5d3a99d5daf79526310144bd8b0a16330be4"} Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.597995 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185"} Oct 08 16:22:49 crc kubenswrapper[4945]: I1008 16:22:49.598017 4945 scope.go:117] "RemoveContainer" containerID="564dc97ba218ea23e76b1887ad38c31e68a3438bd5512f9095fed578a4e1d439" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.268001 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fhn66"] Oct 08 16:23:49 crc kubenswrapper[4945]: E1008 16:23:49.269319 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="registry-server" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.269339 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="registry-server" Oct 08 16:23:49 crc kubenswrapper[4945]: E1008 16:23:49.269367 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="extract-utilities" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.269376 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="extract-utilities" Oct 08 16:23:49 crc kubenswrapper[4945]: E1008 16:23:49.269389 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="extract-content" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.269397 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="extract-content" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.269662 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc50f02-6cbd-462b-9b74-5079b6cb751d" containerName="registry-server" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.271493 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.277134 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fhn66"] Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.456965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-catalog-content\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.457042 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsjd5\" (UniqueName: \"kubernetes.io/projected/9d3e0095-5774-4ba1-9895-604f992db6bd-kube-api-access-lsjd5\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.457202 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-utilities\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.559153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsjd5\" (UniqueName: \"kubernetes.io/projected/9d3e0095-5774-4ba1-9895-604f992db6bd-kube-api-access-lsjd5\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.559328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-utilities\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.559486 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-catalog-content\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.559916 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-utilities\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.560057 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-catalog-content\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.588650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsjd5\" (UniqueName: \"kubernetes.io/projected/9d3e0095-5774-4ba1-9895-604f992db6bd-kube-api-access-lsjd5\") pod \"redhat-operators-fhn66\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:49 crc kubenswrapper[4945]: I1008 16:23:49.594151 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:50 crc kubenswrapper[4945]: I1008 16:23:50.071195 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fhn66"] Oct 08 16:23:50 crc kubenswrapper[4945]: I1008 16:23:50.153454 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fhn66" event={"ID":"9d3e0095-5774-4ba1-9895-604f992db6bd","Type":"ContainerStarted","Data":"e189354fe848f9614f22066f61425de9a14aff73b471df5965a682c2982d30d3"} Oct 08 16:23:51 crc kubenswrapper[4945]: I1008 16:23:51.163370 4945 generic.go:334] "Generic (PLEG): container finished" podID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerID="a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532" exitCode=0 Oct 08 16:23:51 crc kubenswrapper[4945]: I1008 16:23:51.163423 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fhn66" event={"ID":"9d3e0095-5774-4ba1-9895-604f992db6bd","Type":"ContainerDied","Data":"a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532"} Oct 08 16:23:51 crc kubenswrapper[4945]: I1008 16:23:51.165700 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 16:23:53 crc kubenswrapper[4945]: I1008 16:23:53.183608 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fhn66" event={"ID":"9d3e0095-5774-4ba1-9895-604f992db6bd","Type":"ContainerStarted","Data":"b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9"} Oct 08 16:23:57 crc kubenswrapper[4945]: I1008 16:23:57.225660 4945 generic.go:334] "Generic (PLEG): container finished" podID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerID="b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9" exitCode=0 Oct 08 16:23:57 crc kubenswrapper[4945]: I1008 16:23:57.225749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fhn66" event={"ID":"9d3e0095-5774-4ba1-9895-604f992db6bd","Type":"ContainerDied","Data":"b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9"} Oct 08 16:23:59 crc kubenswrapper[4945]: I1008 16:23:59.245226 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fhn66" event={"ID":"9d3e0095-5774-4ba1-9895-604f992db6bd","Type":"ContainerStarted","Data":"2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823"} Oct 08 16:23:59 crc kubenswrapper[4945]: I1008 16:23:59.270580 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fhn66" podStartSLOduration=3.763174485 podStartE2EDuration="10.270559367s" podCreationTimestamp="2025-10-08 16:23:49 +0000 UTC" firstStartedPulling="2025-10-08 16:23:51.16535925 +0000 UTC m=+4240.519274151" lastFinishedPulling="2025-10-08 16:23:57.672744132 +0000 UTC m=+4247.026659033" observedRunningTime="2025-10-08 16:23:59.261856432 +0000 UTC m=+4248.615771333" watchObservedRunningTime="2025-10-08 16:23:59.270559367 +0000 UTC m=+4248.624474268" Oct 08 16:23:59 crc kubenswrapper[4945]: I1008 16:23:59.594554 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:23:59 crc kubenswrapper[4945]: I1008 16:23:59.594762 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:24:00 crc kubenswrapper[4945]: I1008 16:24:00.644677 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fhn66" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="registry-server" probeResult="failure" output=< Oct 08 16:24:00 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:24:00 crc kubenswrapper[4945]: > Oct 08 16:24:10 crc kubenswrapper[4945]: I1008 16:24:10.640332 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fhn66" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="registry-server" probeResult="failure" output=< Oct 08 16:24:10 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:24:10 crc kubenswrapper[4945]: > Oct 08 16:24:19 crc kubenswrapper[4945]: I1008 16:24:19.647781 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:24:19 crc kubenswrapper[4945]: I1008 16:24:19.701420 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:24:20 crc kubenswrapper[4945]: I1008 16:24:20.470276 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fhn66"] Oct 08 16:24:21 crc kubenswrapper[4945]: I1008 16:24:21.460614 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fhn66" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="registry-server" containerID="cri-o://2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823" gracePeriod=2 Oct 08 16:24:21 crc kubenswrapper[4945]: I1008 16:24:21.949166 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.034426 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-utilities\") pod \"9d3e0095-5774-4ba1-9895-604f992db6bd\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.034548 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-catalog-content\") pod \"9d3e0095-5774-4ba1-9895-604f992db6bd\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.034735 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsjd5\" (UniqueName: \"kubernetes.io/projected/9d3e0095-5774-4ba1-9895-604f992db6bd-kube-api-access-lsjd5\") pod \"9d3e0095-5774-4ba1-9895-604f992db6bd\" (UID: \"9d3e0095-5774-4ba1-9895-604f992db6bd\") " Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.035344 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-utilities" (OuterVolumeSpecName: "utilities") pod "9d3e0095-5774-4ba1-9895-604f992db6bd" (UID: "9d3e0095-5774-4ba1-9895-604f992db6bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.040740 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3e0095-5774-4ba1-9895-604f992db6bd-kube-api-access-lsjd5" (OuterVolumeSpecName: "kube-api-access-lsjd5") pod "9d3e0095-5774-4ba1-9895-604f992db6bd" (UID: "9d3e0095-5774-4ba1-9895-604f992db6bd"). InnerVolumeSpecName "kube-api-access-lsjd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.127961 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d3e0095-5774-4ba1-9895-604f992db6bd" (UID: "9d3e0095-5774-4ba1-9895-604f992db6bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.136844 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.136878 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsjd5\" (UniqueName: \"kubernetes.io/projected/9d3e0095-5774-4ba1-9895-604f992db6bd-kube-api-access-lsjd5\") on node \"crc\" DevicePath \"\"" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.136891 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d3e0095-5774-4ba1-9895-604f992db6bd-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.476900 4945 generic.go:334] "Generic (PLEG): container finished" podID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerID="2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823" exitCode=0 Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.476943 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fhn66" event={"ID":"9d3e0095-5774-4ba1-9895-604f992db6bd","Type":"ContainerDied","Data":"2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823"} Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.476975 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fhn66" event={"ID":"9d3e0095-5774-4ba1-9895-604f992db6bd","Type":"ContainerDied","Data":"e189354fe848f9614f22066f61425de9a14aff73b471df5965a682c2982d30d3"} Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.476975 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fhn66" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.477021 4945 scope.go:117] "RemoveContainer" containerID="2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.509064 4945 scope.go:117] "RemoveContainer" containerID="b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.523141 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fhn66"] Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.537263 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fhn66"] Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.544201 4945 scope.go:117] "RemoveContainer" containerID="a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.583297 4945 scope.go:117] "RemoveContainer" containerID="2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823" Oct 08 16:24:22 crc kubenswrapper[4945]: E1008 16:24:22.583851 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823\": container with ID starting with 2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823 not found: ID does not exist" containerID="2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.583891 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823"} err="failed to get container status \"2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823\": rpc error: code = NotFound desc = could not find container \"2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823\": container with ID starting with 2370f8f8de648c0464c50bb3962018872770a80f88425f7e7be628f8e06c0823 not found: ID does not exist" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.583914 4945 scope.go:117] "RemoveContainer" containerID="b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9" Oct 08 16:24:22 crc kubenswrapper[4945]: E1008 16:24:22.584332 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9\": container with ID starting with b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9 not found: ID does not exist" containerID="b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.584363 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9"} err="failed to get container status \"b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9\": rpc error: code = NotFound desc = could not find container \"b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9\": container with ID starting with b25dfb05ff91d3d69117f10e4c556f8212c491a1e22228c95fa0d09186a6e1a9 not found: ID does not exist" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.584381 4945 scope.go:117] "RemoveContainer" containerID="a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532" Oct 08 16:24:22 crc kubenswrapper[4945]: E1008 16:24:22.584750 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532\": container with ID starting with a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532 not found: ID does not exist" containerID="a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532" Oct 08 16:24:22 crc kubenswrapper[4945]: I1008 16:24:22.584775 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532"} err="failed to get container status \"a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532\": rpc error: code = NotFound desc = could not find container \"a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532\": container with ID starting with a124c91cf73790724d0da319d60ef1969c3d0aaa8dcb5ff684f058d167146532 not found: ID does not exist" Oct 08 16:24:24 crc kubenswrapper[4945]: I1008 16:24:24.037984 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" path="/var/lib/kubelet/pods/9d3e0095-5774-4ba1-9895-604f992db6bd/volumes" Oct 08 16:24:49 crc kubenswrapper[4945]: I1008 16:24:49.184228 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:24:49 crc kubenswrapper[4945]: I1008 16:24:49.184839 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:25:19 crc kubenswrapper[4945]: I1008 16:25:19.184541 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:25:19 crc kubenswrapper[4945]: I1008 16:25:19.185146 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.184542 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.185093 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.185166 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.185793 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.185847 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" gracePeriod=600 Oct 08 16:25:49 crc kubenswrapper[4945]: E1008 16:25:49.316168 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.343477 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" exitCode=0 Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.343531 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185"} Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.343601 4945 scope.go:117] "RemoveContainer" containerID="4f31cb254505165039065bc9149b5d3a99d5daf79526310144bd8b0a16330be4" Oct 08 16:25:49 crc kubenswrapper[4945]: I1008 16:25:49.344542 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:25:49 crc kubenswrapper[4945]: E1008 16:25:49.345025 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.644605 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x965s"] Oct 08 16:25:50 crc kubenswrapper[4945]: E1008 16:25:50.645491 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="registry-server" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.645510 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="registry-server" Oct 08 16:25:50 crc kubenswrapper[4945]: E1008 16:25:50.645524 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="extract-content" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.645543 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="extract-content" Oct 08 16:25:50 crc kubenswrapper[4945]: E1008 16:25:50.645573 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="extract-utilities" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.645581 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="extract-utilities" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.645855 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3e0095-5774-4ba1-9895-604f992db6bd" containerName="registry-server" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.647683 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.655928 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x965s"] Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.700786 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-utilities\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.701167 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-catalog-content\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.701335 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m7sd\" (UniqueName: \"kubernetes.io/projected/2bd4e129-301e-438d-b7c3-a994ee884f03-kube-api-access-9m7sd\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.803072 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-utilities\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.803555 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-catalog-content\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.803646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m7sd\" (UniqueName: \"kubernetes.io/projected/2bd4e129-301e-438d-b7c3-a994ee884f03-kube-api-access-9m7sd\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.803919 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-utilities\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.804187 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-catalog-content\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.833219 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m7sd\" (UniqueName: \"kubernetes.io/projected/2bd4e129-301e-438d-b7c3-a994ee884f03-kube-api-access-9m7sd\") pod \"certified-operators-x965s\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:50 crc kubenswrapper[4945]: I1008 16:25:50.972666 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:25:51 crc kubenswrapper[4945]: I1008 16:25:51.497027 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x965s"] Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.371849 4945 generic.go:334] "Generic (PLEG): container finished" podID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerID="75e35342b6dccce6376cb199765260c00b3665beec50ab888c4dd4c83789755b" exitCode=0 Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.371918 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x965s" event={"ID":"2bd4e129-301e-438d-b7c3-a994ee884f03","Type":"ContainerDied","Data":"75e35342b6dccce6376cb199765260c00b3665beec50ab888c4dd4c83789755b"} Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.372152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x965s" event={"ID":"2bd4e129-301e-438d-b7c3-a994ee884f03","Type":"ContainerStarted","Data":"d7fd51408f05afb61789323b90640550a1ee1b24fe1ad7d898071f50daff499e"} Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.442714 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r8c57"] Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.444927 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.455438 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r8c57"] Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.539390 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-utilities\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.539480 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-catalog-content\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.539643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7xsr\" (UniqueName: \"kubernetes.io/projected/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-kube-api-access-v7xsr\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.642303 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-utilities\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.642409 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-catalog-content\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.642503 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7xsr\" (UniqueName: \"kubernetes.io/projected/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-kube-api-access-v7xsr\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.642980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-utilities\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.643000 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-catalog-content\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.672998 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7xsr\" (UniqueName: \"kubernetes.io/projected/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-kube-api-access-v7xsr\") pod \"redhat-marketplace-r8c57\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:52 crc kubenswrapper[4945]: I1008 16:25:52.775210 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.051155 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pvxkh"] Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.054474 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.062920 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pvxkh"] Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.156192 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-catalog-content\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.156729 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-utilities\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.158364 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvvrx\" (UniqueName: \"kubernetes.io/projected/3c616c50-d104-4fdb-9e3b-521126b4227d-kube-api-access-vvvrx\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.260142 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvvrx\" (UniqueName: \"kubernetes.io/projected/3c616c50-d104-4fdb-9e3b-521126b4227d-kube-api-access-vvvrx\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.260451 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-catalog-content\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.260517 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-utilities\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.261263 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-catalog-content\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.261276 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-utilities\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.277235 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r8c57"] Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.285718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvvrx\" (UniqueName: \"kubernetes.io/projected/3c616c50-d104-4fdb-9e3b-521126b4227d-kube-api-access-vvvrx\") pod \"community-operators-pvxkh\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.381503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r8c57" event={"ID":"fd6d3e58-5669-4ee9-ae12-6563f09c2a89","Type":"ContainerStarted","Data":"c0dbfc015693b30be3c7677f334f4bc04f06f5be698f5ec0c1cdda85b793e6ca"} Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.386238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:25:53 crc kubenswrapper[4945]: I1008 16:25:53.896410 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pvxkh"] Oct 08 16:25:54 crc kubenswrapper[4945]: I1008 16:25:54.392871 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerID="329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f" exitCode=0 Oct 08 16:25:54 crc kubenswrapper[4945]: I1008 16:25:54.393184 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r8c57" event={"ID":"fd6d3e58-5669-4ee9-ae12-6563f09c2a89","Type":"ContainerDied","Data":"329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f"} Oct 08 16:25:54 crc kubenswrapper[4945]: I1008 16:25:54.401168 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerID="f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d" exitCode=0 Oct 08 16:25:54 crc kubenswrapper[4945]: I1008 16:25:54.401249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvxkh" event={"ID":"3c616c50-d104-4fdb-9e3b-521126b4227d","Type":"ContainerDied","Data":"f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d"} Oct 08 16:25:54 crc kubenswrapper[4945]: I1008 16:25:54.401282 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvxkh" event={"ID":"3c616c50-d104-4fdb-9e3b-521126b4227d","Type":"ContainerStarted","Data":"2096353687bf459211ec8795be351e4f43519ad7147adb6a4c04c5ecb67da5b5"} Oct 08 16:25:54 crc kubenswrapper[4945]: I1008 16:25:54.406958 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x965s" event={"ID":"2bd4e129-301e-438d-b7c3-a994ee884f03","Type":"ContainerStarted","Data":"34a0198770badf00fc4235e5af0f2428bcbabb2eba51181596085aa2bd9ac7be"} Oct 08 16:25:55 crc kubenswrapper[4945]: I1008 16:25:55.419985 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvxkh" event={"ID":"3c616c50-d104-4fdb-9e3b-521126b4227d","Type":"ContainerStarted","Data":"c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da"} Oct 08 16:25:55 crc kubenswrapper[4945]: I1008 16:25:55.423695 4945 generic.go:334] "Generic (PLEG): container finished" podID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerID="34a0198770badf00fc4235e5af0f2428bcbabb2eba51181596085aa2bd9ac7be" exitCode=0 Oct 08 16:25:55 crc kubenswrapper[4945]: I1008 16:25:55.423741 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x965s" event={"ID":"2bd4e129-301e-438d-b7c3-a994ee884f03","Type":"ContainerDied","Data":"34a0198770badf00fc4235e5af0f2428bcbabb2eba51181596085aa2bd9ac7be"} Oct 08 16:25:56 crc kubenswrapper[4945]: I1008 16:25:56.435495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x965s" event={"ID":"2bd4e129-301e-438d-b7c3-a994ee884f03","Type":"ContainerStarted","Data":"37ff214562922ffe2e843cc78dfe5a4d700c9f478b121456282f779971e17a41"} Oct 08 16:25:56 crc kubenswrapper[4945]: I1008 16:25:56.438456 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerID="38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1" exitCode=0 Oct 08 16:25:56 crc kubenswrapper[4945]: I1008 16:25:56.438591 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r8c57" event={"ID":"fd6d3e58-5669-4ee9-ae12-6563f09c2a89","Type":"ContainerDied","Data":"38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1"} Oct 08 16:25:56 crc kubenswrapper[4945]: I1008 16:25:56.464473 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x965s" podStartSLOduration=2.9266321299999998 podStartE2EDuration="6.464452687s" podCreationTimestamp="2025-10-08 16:25:50 +0000 UTC" firstStartedPulling="2025-10-08 16:25:52.373340522 +0000 UTC m=+4361.727255423" lastFinishedPulling="2025-10-08 16:25:55.911161079 +0000 UTC m=+4365.265075980" observedRunningTime="2025-10-08 16:25:56.457873123 +0000 UTC m=+4365.811788034" watchObservedRunningTime="2025-10-08 16:25:56.464452687 +0000 UTC m=+4365.818367588" Oct 08 16:25:58 crc kubenswrapper[4945]: I1008 16:25:58.468417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r8c57" event={"ID":"fd6d3e58-5669-4ee9-ae12-6563f09c2a89","Type":"ContainerStarted","Data":"eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a"} Oct 08 16:25:58 crc kubenswrapper[4945]: I1008 16:25:58.472235 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerID="c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da" exitCode=0 Oct 08 16:25:58 crc kubenswrapper[4945]: I1008 16:25:58.472291 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvxkh" event={"ID":"3c616c50-d104-4fdb-9e3b-521126b4227d","Type":"ContainerDied","Data":"c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da"} Oct 08 16:25:58 crc kubenswrapper[4945]: I1008 16:25:58.497640 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r8c57" podStartSLOduration=3.777064474 podStartE2EDuration="6.497619866s" podCreationTimestamp="2025-10-08 16:25:52 +0000 UTC" firstStartedPulling="2025-10-08 16:25:54.394859344 +0000 UTC m=+4363.748774265" lastFinishedPulling="2025-10-08 16:25:57.115414756 +0000 UTC m=+4366.469329657" observedRunningTime="2025-10-08 16:25:58.494049887 +0000 UTC m=+4367.847964788" watchObservedRunningTime="2025-10-08 16:25:58.497619866 +0000 UTC m=+4367.851534767" Oct 08 16:25:59 crc kubenswrapper[4945]: I1008 16:25:59.484422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvxkh" event={"ID":"3c616c50-d104-4fdb-9e3b-521126b4227d","Type":"ContainerStarted","Data":"2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277"} Oct 08 16:25:59 crc kubenswrapper[4945]: I1008 16:25:59.509353 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pvxkh" podStartSLOduration=1.8556077690000001 podStartE2EDuration="6.509331681s" podCreationTimestamp="2025-10-08 16:25:53 +0000 UTC" firstStartedPulling="2025-10-08 16:25:54.404624256 +0000 UTC m=+4363.758539157" lastFinishedPulling="2025-10-08 16:25:59.058348158 +0000 UTC m=+4368.412263069" observedRunningTime="2025-10-08 16:25:59.502916372 +0000 UTC m=+4368.856831273" watchObservedRunningTime="2025-10-08 16:25:59.509331681 +0000 UTC m=+4368.863246582" Oct 08 16:26:00 crc kubenswrapper[4945]: I1008 16:26:00.973390 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:26:00 crc kubenswrapper[4945]: I1008 16:26:00.973779 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:26:01 crc kubenswrapper[4945]: I1008 16:26:01.024916 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:26:01 crc kubenswrapper[4945]: E1008 16:26:01.025432 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:26:01 crc kubenswrapper[4945]: I1008 16:26:01.026091 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:26:01 crc kubenswrapper[4945]: I1008 16:26:01.568707 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:26:02 crc kubenswrapper[4945]: I1008 16:26:02.775795 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:26:02 crc kubenswrapper[4945]: I1008 16:26:02.776191 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:26:02 crc kubenswrapper[4945]: I1008 16:26:02.912766 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:26:03 crc kubenswrapper[4945]: I1008 16:26:03.386830 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:26:03 crc kubenswrapper[4945]: I1008 16:26:03.387146 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:26:03 crc kubenswrapper[4945]: I1008 16:26:03.572070 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:26:03 crc kubenswrapper[4945]: I1008 16:26:03.636083 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x965s"] Oct 08 16:26:03 crc kubenswrapper[4945]: I1008 16:26:03.636390 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x965s" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="registry-server" containerID="cri-o://37ff214562922ffe2e843cc78dfe5a4d700c9f478b121456282f779971e17a41" gracePeriod=2 Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.434185 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-pvxkh" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="registry-server" probeResult="failure" output=< Oct 08 16:26:04 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:26:04 crc kubenswrapper[4945]: > Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.534084 4945 generic.go:334] "Generic (PLEG): container finished" podID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerID="37ff214562922ffe2e843cc78dfe5a4d700c9f478b121456282f779971e17a41" exitCode=0 Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.535123 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x965s" event={"ID":"2bd4e129-301e-438d-b7c3-a994ee884f03","Type":"ContainerDied","Data":"37ff214562922ffe2e843cc78dfe5a4d700c9f478b121456282f779971e17a41"} Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.754522 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.853281 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-catalog-content\") pod \"2bd4e129-301e-438d-b7c3-a994ee884f03\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.853372 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m7sd\" (UniqueName: \"kubernetes.io/projected/2bd4e129-301e-438d-b7c3-a994ee884f03-kube-api-access-9m7sd\") pod \"2bd4e129-301e-438d-b7c3-a994ee884f03\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.853574 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-utilities\") pod \"2bd4e129-301e-438d-b7c3-a994ee884f03\" (UID: \"2bd4e129-301e-438d-b7c3-a994ee884f03\") " Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.854435 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-utilities" (OuterVolumeSpecName: "utilities") pod "2bd4e129-301e-438d-b7c3-a994ee884f03" (UID: "2bd4e129-301e-438d-b7c3-a994ee884f03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.859636 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd4e129-301e-438d-b7c3-a994ee884f03-kube-api-access-9m7sd" (OuterVolumeSpecName: "kube-api-access-9m7sd") pod "2bd4e129-301e-438d-b7c3-a994ee884f03" (UID: "2bd4e129-301e-438d-b7c3-a994ee884f03"). InnerVolumeSpecName "kube-api-access-9m7sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.894240 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bd4e129-301e-438d-b7c3-a994ee884f03" (UID: "2bd4e129-301e-438d-b7c3-a994ee884f03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.955896 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m7sd\" (UniqueName: \"kubernetes.io/projected/2bd4e129-301e-438d-b7c3-a994ee884f03-kube-api-access-9m7sd\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.955947 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:04 crc kubenswrapper[4945]: I1008 16:26:04.955964 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd4e129-301e-438d-b7c3-a994ee884f03-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.545322 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x965s" event={"ID":"2bd4e129-301e-438d-b7c3-a994ee884f03","Type":"ContainerDied","Data":"d7fd51408f05afb61789323b90640550a1ee1b24fe1ad7d898071f50daff499e"} Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.545371 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x965s" Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.545378 4945 scope.go:117] "RemoveContainer" containerID="37ff214562922ffe2e843cc78dfe5a4d700c9f478b121456282f779971e17a41" Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.575211 4945 scope.go:117] "RemoveContainer" containerID="34a0198770badf00fc4235e5af0f2428bcbabb2eba51181596085aa2bd9ac7be" Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.587050 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x965s"] Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.607887 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x965s"] Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.614040 4945 scope.go:117] "RemoveContainer" containerID="75e35342b6dccce6376cb199765260c00b3665beec50ab888c4dd4c83789755b" Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.843528 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r8c57"] Oct 08 16:26:05 crc kubenswrapper[4945]: I1008 16:26:05.843787 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r8c57" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="registry-server" containerID="cri-o://eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a" gracePeriod=2 Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.046209 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" path="/var/lib/kubelet/pods/2bd4e129-301e-438d-b7c3-a994ee884f03/volumes" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.445930 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.558827 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerID="eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a" exitCode=0 Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.558865 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r8c57" event={"ID":"fd6d3e58-5669-4ee9-ae12-6563f09c2a89","Type":"ContainerDied","Data":"eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a"} Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.558893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r8c57" event={"ID":"fd6d3e58-5669-4ee9-ae12-6563f09c2a89","Type":"ContainerDied","Data":"c0dbfc015693b30be3c7677f334f4bc04f06f5be698f5ec0c1cdda85b793e6ca"} Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.558909 4945 scope.go:117] "RemoveContainer" containerID="eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.558873 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r8c57" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.579759 4945 scope.go:117] "RemoveContainer" containerID="38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.591345 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-catalog-content\") pod \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.591420 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7xsr\" (UniqueName: \"kubernetes.io/projected/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-kube-api-access-v7xsr\") pod \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.591587 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-utilities\") pod \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\" (UID: \"fd6d3e58-5669-4ee9-ae12-6563f09c2a89\") " Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.592625 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-utilities" (OuterVolumeSpecName: "utilities") pod "fd6d3e58-5669-4ee9-ae12-6563f09c2a89" (UID: "fd6d3e58-5669-4ee9-ae12-6563f09c2a89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.598292 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-kube-api-access-v7xsr" (OuterVolumeSpecName: "kube-api-access-v7xsr") pod "fd6d3e58-5669-4ee9-ae12-6563f09c2a89" (UID: "fd6d3e58-5669-4ee9-ae12-6563f09c2a89"). InnerVolumeSpecName "kube-api-access-v7xsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.601734 4945 scope.go:117] "RemoveContainer" containerID="329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.613949 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd6d3e58-5669-4ee9-ae12-6563f09c2a89" (UID: "fd6d3e58-5669-4ee9-ae12-6563f09c2a89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.693293 4945 scope.go:117] "RemoveContainer" containerID="eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a" Oct 08 16:26:06 crc kubenswrapper[4945]: E1008 16:26:06.693649 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a\": container with ID starting with eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a not found: ID does not exist" containerID="eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.693690 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.693724 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7xsr\" (UniqueName: \"kubernetes.io/projected/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-kube-api-access-v7xsr\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.693741 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6d3e58-5669-4ee9-ae12-6563f09c2a89-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.693681 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a"} err="failed to get container status \"eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a\": rpc error: code = NotFound desc = could not find container \"eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a\": container with ID starting with eebfc9b95540e0dbd3cae3604a01de0ff172991dc14937e09928ac0ced1c404a not found: ID does not exist" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.693760 4945 scope.go:117] "RemoveContainer" containerID="38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1" Oct 08 16:26:06 crc kubenswrapper[4945]: E1008 16:26:06.694001 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1\": container with ID starting with 38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1 not found: ID does not exist" containerID="38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.694036 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1"} err="failed to get container status \"38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1\": rpc error: code = NotFound desc = could not find container \"38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1\": container with ID starting with 38eefbfec2729a1e42d3057c225387a01e2ee6a5921113e4a53e58e1d727eaf1 not found: ID does not exist" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.694056 4945 scope.go:117] "RemoveContainer" containerID="329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f" Oct 08 16:26:06 crc kubenswrapper[4945]: E1008 16:26:06.694442 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f\": container with ID starting with 329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f not found: ID does not exist" containerID="329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.694479 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f"} err="failed to get container status \"329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f\": rpc error: code = NotFound desc = could not find container \"329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f\": container with ID starting with 329e1ee0d737a7e6f3fcccd8f251355ea6d3809b19395c1835d5da8b7ea14e7f not found: ID does not exist" Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.904474 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r8c57"] Oct 08 16:26:06 crc kubenswrapper[4945]: I1008 16:26:06.913214 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r8c57"] Oct 08 16:26:08 crc kubenswrapper[4945]: I1008 16:26:08.036804 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" path="/var/lib/kubelet/pods/fd6d3e58-5669-4ee9-ae12-6563f09c2a89/volumes" Oct 08 16:26:13 crc kubenswrapper[4945]: I1008 16:26:13.025667 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:26:13 crc kubenswrapper[4945]: E1008 16:26:13.026854 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:26:13 crc kubenswrapper[4945]: I1008 16:26:13.433855 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:26:13 crc kubenswrapper[4945]: I1008 16:26:13.480748 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:26:13 crc kubenswrapper[4945]: I1008 16:26:13.669831 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pvxkh"] Oct 08 16:26:14 crc kubenswrapper[4945]: I1008 16:26:14.631950 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pvxkh" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="registry-server" containerID="cri-o://2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277" gracePeriod=2 Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.247229 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.384388 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvvrx\" (UniqueName: \"kubernetes.io/projected/3c616c50-d104-4fdb-9e3b-521126b4227d-kube-api-access-vvvrx\") pod \"3c616c50-d104-4fdb-9e3b-521126b4227d\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.384478 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-catalog-content\") pod \"3c616c50-d104-4fdb-9e3b-521126b4227d\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.384512 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-utilities\") pod \"3c616c50-d104-4fdb-9e3b-521126b4227d\" (UID: \"3c616c50-d104-4fdb-9e3b-521126b4227d\") " Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.385548 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-utilities" (OuterVolumeSpecName: "utilities") pod "3c616c50-d104-4fdb-9e3b-521126b4227d" (UID: "3c616c50-d104-4fdb-9e3b-521126b4227d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.390798 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c616c50-d104-4fdb-9e3b-521126b4227d-kube-api-access-vvvrx" (OuterVolumeSpecName: "kube-api-access-vvvrx") pod "3c616c50-d104-4fdb-9e3b-521126b4227d" (UID: "3c616c50-d104-4fdb-9e3b-521126b4227d"). InnerVolumeSpecName "kube-api-access-vvvrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.433441 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c616c50-d104-4fdb-9e3b-521126b4227d" (UID: "3c616c50-d104-4fdb-9e3b-521126b4227d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.487292 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvvrx\" (UniqueName: \"kubernetes.io/projected/3c616c50-d104-4fdb-9e3b-521126b4227d-kube-api-access-vvvrx\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.487331 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.487340 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c616c50-d104-4fdb-9e3b-521126b4227d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.650077 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerID="2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277" exitCode=0 Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.650252 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvxkh" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.650277 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvxkh" event={"ID":"3c616c50-d104-4fdb-9e3b-521126b4227d","Type":"ContainerDied","Data":"2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277"} Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.650417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvxkh" event={"ID":"3c616c50-d104-4fdb-9e3b-521126b4227d","Type":"ContainerDied","Data":"2096353687bf459211ec8795be351e4f43519ad7147adb6a4c04c5ecb67da5b5"} Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.650440 4945 scope.go:117] "RemoveContainer" containerID="2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.674140 4945 scope.go:117] "RemoveContainer" containerID="c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.685936 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pvxkh"] Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.694008 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pvxkh"] Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.704237 4945 scope.go:117] "RemoveContainer" containerID="f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.747462 4945 scope.go:117] "RemoveContainer" containerID="2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277" Oct 08 16:26:15 crc kubenswrapper[4945]: E1008 16:26:15.747865 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277\": container with ID starting with 2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277 not found: ID does not exist" containerID="2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.747897 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277"} err="failed to get container status \"2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277\": rpc error: code = NotFound desc = could not find container \"2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277\": container with ID starting with 2d44865884663cacf9de5037cbd345067ab2b68ec9ef26e89dc7d717c0bb9277 not found: ID does not exist" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.747920 4945 scope.go:117] "RemoveContainer" containerID="c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da" Oct 08 16:26:15 crc kubenswrapper[4945]: E1008 16:26:15.748301 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da\": container with ID starting with c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da not found: ID does not exist" containerID="c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.748343 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da"} err="failed to get container status \"c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da\": rpc error: code = NotFound desc = could not find container \"c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da\": container with ID starting with c4736f5b8a0b1dd4d0696c412c2aa06412dfbb48effc85a1f45fd85810f539da not found: ID does not exist" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.748371 4945 scope.go:117] "RemoveContainer" containerID="f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d" Oct 08 16:26:15 crc kubenswrapper[4945]: E1008 16:26:15.748868 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d\": container with ID starting with f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d not found: ID does not exist" containerID="f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d" Oct 08 16:26:15 crc kubenswrapper[4945]: I1008 16:26:15.748898 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d"} err="failed to get container status \"f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d\": rpc error: code = NotFound desc = could not find container \"f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d\": container with ID starting with f8429dc24233e2d700f912c745b33634ec2d40230d8674dc90ca5a3b53c3556d not found: ID does not exist" Oct 08 16:26:16 crc kubenswrapper[4945]: I1008 16:26:16.036364 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" path="/var/lib/kubelet/pods/3c616c50-d104-4fdb-9e3b-521126b4227d/volumes" Oct 08 16:26:24 crc kubenswrapper[4945]: I1008 16:26:24.025162 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:26:24 crc kubenswrapper[4945]: E1008 16:26:24.026625 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:26:39 crc kubenswrapper[4945]: I1008 16:26:39.024605 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:26:39 crc kubenswrapper[4945]: E1008 16:26:39.025393 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:26:50 crc kubenswrapper[4945]: I1008 16:26:50.023849 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:26:50 crc kubenswrapper[4945]: E1008 16:26:50.024620 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:27:01 crc kubenswrapper[4945]: I1008 16:27:01.023957 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:27:01 crc kubenswrapper[4945]: E1008 16:27:01.024786 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:27:16 crc kubenswrapper[4945]: I1008 16:27:16.026565 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:27:16 crc kubenswrapper[4945]: E1008 16:27:16.028030 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:27:31 crc kubenswrapper[4945]: I1008 16:27:31.024151 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:27:31 crc kubenswrapper[4945]: E1008 16:27:31.025184 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:27:46 crc kubenswrapper[4945]: I1008 16:27:46.025408 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:27:46 crc kubenswrapper[4945]: E1008 16:27:46.026187 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:27:58 crc kubenswrapper[4945]: I1008 16:27:58.029431 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:27:58 crc kubenswrapper[4945]: E1008 16:27:58.039959 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:28:11 crc kubenswrapper[4945]: I1008 16:28:11.024331 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:28:11 crc kubenswrapper[4945]: E1008 16:28:11.025289 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:28:22 crc kubenswrapper[4945]: I1008 16:28:22.031392 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:28:22 crc kubenswrapper[4945]: E1008 16:28:22.032191 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:28:35 crc kubenswrapper[4945]: I1008 16:28:35.024649 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:28:35 crc kubenswrapper[4945]: E1008 16:28:35.025408 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:28:46 crc kubenswrapper[4945]: I1008 16:28:46.024493 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:28:46 crc kubenswrapper[4945]: E1008 16:28:46.026227 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:29:00 crc kubenswrapper[4945]: I1008 16:29:00.024925 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:29:00 crc kubenswrapper[4945]: E1008 16:29:00.025676 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:29:13 crc kubenswrapper[4945]: I1008 16:29:13.024194 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:29:13 crc kubenswrapper[4945]: E1008 16:29:13.025399 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:29:26 crc kubenswrapper[4945]: I1008 16:29:26.024732 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:29:26 crc kubenswrapper[4945]: E1008 16:29:26.025611 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:29:41 crc kubenswrapper[4945]: I1008 16:29:41.024253 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:29:41 crc kubenswrapper[4945]: E1008 16:29:41.025026 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:29:56 crc kubenswrapper[4945]: I1008 16:29:56.024194 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:29:56 crc kubenswrapper[4945]: E1008 16:29:56.024972 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.172135 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp"] Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.173824 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="extract-utilities" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.173857 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="extract-utilities" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.173886 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="extract-utilities" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.173896 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="extract-utilities" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.173909 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="extract-content" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.173920 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="extract-content" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.173938 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.173945 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.173967 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="extract-utilities" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.173976 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="extract-utilities" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.174012 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="extract-content" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.174022 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="extract-content" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.174035 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.174043 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.174080 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="extract-content" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.174089 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="extract-content" Oct 08 16:30:00 crc kubenswrapper[4945]: E1008 16:30:00.174103 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.174131 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.174506 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c616c50-d104-4fdb-9e3b-521126b4227d" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.174531 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6d3e58-5669-4ee9-ae12-6563f09c2a89" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.174567 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd4e129-301e-438d-b7c3-a994ee884f03" containerName="registry-server" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.176285 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.183026 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.183038 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.188969 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp"] Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.274609 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/682b10fc-5ab4-4a72-80d6-1c71f995e62b-secret-volume\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.274683 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tdw5\" (UniqueName: \"kubernetes.io/projected/682b10fc-5ab4-4a72-80d6-1c71f995e62b-kube-api-access-5tdw5\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.274849 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/682b10fc-5ab4-4a72-80d6-1c71f995e62b-config-volume\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.376716 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/682b10fc-5ab4-4a72-80d6-1c71f995e62b-config-volume\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.376826 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/682b10fc-5ab4-4a72-80d6-1c71f995e62b-secret-volume\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.376854 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tdw5\" (UniqueName: \"kubernetes.io/projected/682b10fc-5ab4-4a72-80d6-1c71f995e62b-kube-api-access-5tdw5\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.378137 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/682b10fc-5ab4-4a72-80d6-1c71f995e62b-config-volume\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.384295 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/682b10fc-5ab4-4a72-80d6-1c71f995e62b-secret-volume\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.398491 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tdw5\" (UniqueName: \"kubernetes.io/projected/682b10fc-5ab4-4a72-80d6-1c71f995e62b-kube-api-access-5tdw5\") pod \"collect-profiles-29332350-xjlmp\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.504527 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:00 crc kubenswrapper[4945]: I1008 16:30:00.933487 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp"] Oct 08 16:30:01 crc kubenswrapper[4945]: I1008 16:30:01.725381 4945 generic.go:334] "Generic (PLEG): container finished" podID="682b10fc-5ab4-4a72-80d6-1c71f995e62b" containerID="51248083fd3f8cd58daf651f3b000eb7d898e214739b439b8ed97af2293b3cb8" exitCode=0 Oct 08 16:30:01 crc kubenswrapper[4945]: I1008 16:30:01.725444 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" event={"ID":"682b10fc-5ab4-4a72-80d6-1c71f995e62b","Type":"ContainerDied","Data":"51248083fd3f8cd58daf651f3b000eb7d898e214739b439b8ed97af2293b3cb8"} Oct 08 16:30:01 crc kubenswrapper[4945]: I1008 16:30:01.725681 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" event={"ID":"682b10fc-5ab4-4a72-80d6-1c71f995e62b","Type":"ContainerStarted","Data":"3341bd3f6c95ad4b96e080fba8cf228082b5994bdb03bb09bb9f994e3e740ac4"} Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.085800 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.232490 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tdw5\" (UniqueName: \"kubernetes.io/projected/682b10fc-5ab4-4a72-80d6-1c71f995e62b-kube-api-access-5tdw5\") pod \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.232641 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/682b10fc-5ab4-4a72-80d6-1c71f995e62b-config-volume\") pod \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.232731 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/682b10fc-5ab4-4a72-80d6-1c71f995e62b-secret-volume\") pod \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\" (UID: \"682b10fc-5ab4-4a72-80d6-1c71f995e62b\") " Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.233342 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/682b10fc-5ab4-4a72-80d6-1c71f995e62b-config-volume" (OuterVolumeSpecName: "config-volume") pod "682b10fc-5ab4-4a72-80d6-1c71f995e62b" (UID: "682b10fc-5ab4-4a72-80d6-1c71f995e62b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.238506 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/682b10fc-5ab4-4a72-80d6-1c71f995e62b-kube-api-access-5tdw5" (OuterVolumeSpecName: "kube-api-access-5tdw5") pod "682b10fc-5ab4-4a72-80d6-1c71f995e62b" (UID: "682b10fc-5ab4-4a72-80d6-1c71f995e62b"). InnerVolumeSpecName "kube-api-access-5tdw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.238526 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/682b10fc-5ab4-4a72-80d6-1c71f995e62b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "682b10fc-5ab4-4a72-80d6-1c71f995e62b" (UID: "682b10fc-5ab4-4a72-80d6-1c71f995e62b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.335289 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/682b10fc-5ab4-4a72-80d6-1c71f995e62b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.335351 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tdw5\" (UniqueName: \"kubernetes.io/projected/682b10fc-5ab4-4a72-80d6-1c71f995e62b-kube-api-access-5tdw5\") on node \"crc\" DevicePath \"\"" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.335364 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/682b10fc-5ab4-4a72-80d6-1c71f995e62b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.743693 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" event={"ID":"682b10fc-5ab4-4a72-80d6-1c71f995e62b","Type":"ContainerDied","Data":"3341bd3f6c95ad4b96e080fba8cf228082b5994bdb03bb09bb9f994e3e740ac4"} Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.744039 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3341bd3f6c95ad4b96e080fba8cf228082b5994bdb03bb09bb9f994e3e740ac4" Oct 08 16:30:03 crc kubenswrapper[4945]: I1008 16:30:03.743997 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332350-xjlmp" Oct 08 16:30:04 crc kubenswrapper[4945]: I1008 16:30:04.153419 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s"] Oct 08 16:30:04 crc kubenswrapper[4945]: I1008 16:30:04.160793 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-5fj4s"] Oct 08 16:30:06 crc kubenswrapper[4945]: I1008 16:30:06.037226 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7ea7eab-c495-4da0-bdaa-18edd47f0744" path="/var/lib/kubelet/pods/b7ea7eab-c495-4da0-bdaa-18edd47f0744/volumes" Oct 08 16:30:11 crc kubenswrapper[4945]: I1008 16:30:11.024367 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:30:11 crc kubenswrapper[4945]: E1008 16:30:11.025164 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:30:25 crc kubenswrapper[4945]: I1008 16:30:25.024487 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:30:25 crc kubenswrapper[4945]: E1008 16:30:25.025543 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:30:38 crc kubenswrapper[4945]: I1008 16:30:38.024070 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:30:38 crc kubenswrapper[4945]: E1008 16:30:38.024901 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:30:46 crc kubenswrapper[4945]: I1008 16:30:46.650325 4945 scope.go:117] "RemoveContainer" containerID="5620c65ea05b581ef32a4dc80bd7defcda5a705f034c56daf6f3abdca4024dbf" Oct 08 16:30:52 crc kubenswrapper[4945]: I1008 16:30:52.030366 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:30:53 crc kubenswrapper[4945]: I1008 16:30:53.202669 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"2a3055c35e39f4c78367daaba70caaada0e749323c9fd679bc600a59fa8a8233"} Oct 08 16:33:07 crc kubenswrapper[4945]: E1008 16:33:07.334379 4945 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.138:34496->38.102.83.138:44411: write tcp 38.102.83.138:34496->38.102.83.138:44411: write: broken pipe Oct 08 16:33:19 crc kubenswrapper[4945]: I1008 16:33:19.184043 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:33:19 crc kubenswrapper[4945]: I1008 16:33:19.185657 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:33:49 crc kubenswrapper[4945]: I1008 16:33:49.184639 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:33:49 crc kubenswrapper[4945]: I1008 16:33:49.185217 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:34:19 crc kubenswrapper[4945]: I1008 16:34:19.183843 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:34:19 crc kubenswrapper[4945]: I1008 16:34:19.184498 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:34:19 crc kubenswrapper[4945]: I1008 16:34:19.184559 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:34:19 crc kubenswrapper[4945]: I1008 16:34:19.185484 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a3055c35e39f4c78367daaba70caaada0e749323c9fd679bc600a59fa8a8233"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:34:19 crc kubenswrapper[4945]: I1008 16:34:19.185548 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://2a3055c35e39f4c78367daaba70caaada0e749323c9fd679bc600a59fa8a8233" gracePeriod=600 Oct 08 16:34:20 crc kubenswrapper[4945]: I1008 16:34:20.272503 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="2a3055c35e39f4c78367daaba70caaada0e749323c9fd679bc600a59fa8a8233" exitCode=0 Oct 08 16:34:20 crc kubenswrapper[4945]: I1008 16:34:20.272618 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"2a3055c35e39f4c78367daaba70caaada0e749323c9fd679bc600a59fa8a8233"} Oct 08 16:34:20 crc kubenswrapper[4945]: I1008 16:34:20.273146 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4"} Oct 08 16:34:20 crc kubenswrapper[4945]: I1008 16:34:20.273169 4945 scope.go:117] "RemoveContainer" containerID="af2807502398c6d2d15b48b5994e668e3665df15b46844c5086101ff3be59185" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.294124 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hmqtn"] Oct 08 16:35:34 crc kubenswrapper[4945]: E1008 16:35:34.295254 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="682b10fc-5ab4-4a72-80d6-1c71f995e62b" containerName="collect-profiles" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.295270 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="682b10fc-5ab4-4a72-80d6-1c71f995e62b" containerName="collect-profiles" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.295523 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="682b10fc-5ab4-4a72-80d6-1c71f995e62b" containerName="collect-profiles" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.296980 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.306084 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmqtn"] Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.399701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t4wn\" (UniqueName: \"kubernetes.io/projected/cb793be2-9d58-4a3b-814f-17efa4392c96-kube-api-access-2t4wn\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.400213 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-catalog-content\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.400353 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-utilities\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.502594 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-catalog-content\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.502677 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-utilities\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.502714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t4wn\" (UniqueName: \"kubernetes.io/projected/cb793be2-9d58-4a3b-814f-17efa4392c96-kube-api-access-2t4wn\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.503207 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-catalog-content\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:34 crc kubenswrapper[4945]: I1008 16:35:34.503241 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-utilities\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:35 crc kubenswrapper[4945]: I1008 16:35:35.013133 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t4wn\" (UniqueName: \"kubernetes.io/projected/cb793be2-9d58-4a3b-814f-17efa4392c96-kube-api-access-2t4wn\") pod \"redhat-operators-hmqtn\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:35 crc kubenswrapper[4945]: I1008 16:35:35.219672 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:35 crc kubenswrapper[4945]: I1008 16:35:35.823721 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmqtn"] Oct 08 16:35:36 crc kubenswrapper[4945]: I1008 16:35:36.039594 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmqtn" event={"ID":"cb793be2-9d58-4a3b-814f-17efa4392c96","Type":"ContainerStarted","Data":"b2692a3254d508e8dd2f37222237c7f4c2af1a4db181bad1189a7a5d08eca36c"} Oct 08 16:35:37 crc kubenswrapper[4945]: I1008 16:35:37.039907 4945 generic.go:334] "Generic (PLEG): container finished" podID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerID="452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09" exitCode=0 Oct 08 16:35:37 crc kubenswrapper[4945]: I1008 16:35:37.040186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmqtn" event={"ID":"cb793be2-9d58-4a3b-814f-17efa4392c96","Type":"ContainerDied","Data":"452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09"} Oct 08 16:35:37 crc kubenswrapper[4945]: I1008 16:35:37.043237 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 16:35:39 crc kubenswrapper[4945]: I1008 16:35:39.061403 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmqtn" event={"ID":"cb793be2-9d58-4a3b-814f-17efa4392c96","Type":"ContainerStarted","Data":"722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f"} Oct 08 16:35:47 crc kubenswrapper[4945]: I1008 16:35:47.148085 4945 generic.go:334] "Generic (PLEG): container finished" podID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerID="722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f" exitCode=0 Oct 08 16:35:47 crc kubenswrapper[4945]: I1008 16:35:47.148152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmqtn" event={"ID":"cb793be2-9d58-4a3b-814f-17efa4392c96","Type":"ContainerDied","Data":"722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f"} Oct 08 16:35:49 crc kubenswrapper[4945]: I1008 16:35:49.216643 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmqtn" event={"ID":"cb793be2-9d58-4a3b-814f-17efa4392c96","Type":"ContainerStarted","Data":"f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067"} Oct 08 16:35:49 crc kubenswrapper[4945]: I1008 16:35:49.244436 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hmqtn" podStartSLOduration=4.626622458 podStartE2EDuration="15.244415336s" podCreationTimestamp="2025-10-08 16:35:34 +0000 UTC" firstStartedPulling="2025-10-08 16:35:37.042931278 +0000 UTC m=+4946.396846189" lastFinishedPulling="2025-10-08 16:35:47.660724166 +0000 UTC m=+4957.014639067" observedRunningTime="2025-10-08 16:35:49.236174141 +0000 UTC m=+4958.590089042" watchObservedRunningTime="2025-10-08 16:35:49.244415336 +0000 UTC m=+4958.598330237" Oct 08 16:35:55 crc kubenswrapper[4945]: I1008 16:35:55.220450 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:55 crc kubenswrapper[4945]: I1008 16:35:55.221021 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:35:56 crc kubenswrapper[4945]: I1008 16:35:56.273255 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hmqtn" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="registry-server" probeResult="failure" output=< Oct 08 16:35:56 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:35:56 crc kubenswrapper[4945]: > Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.163468 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mrhnc"] Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.166072 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.180132 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrhnc"] Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.354229 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6pzj\" (UniqueName: \"kubernetes.io/projected/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-kube-api-access-d6pzj\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.354333 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-catalog-content\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.354938 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-utilities\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.456863 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6pzj\" (UniqueName: \"kubernetes.io/projected/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-kube-api-access-d6pzj\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.456997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-catalog-content\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.457037 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-utilities\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.457560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-catalog-content\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.457665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-utilities\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.481053 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6pzj\" (UniqueName: \"kubernetes.io/projected/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-kube-api-access-d6pzj\") pod \"certified-operators-mrhnc\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:01 crc kubenswrapper[4945]: I1008 16:36:01.538301 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:02 crc kubenswrapper[4945]: I1008 16:36:02.118536 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrhnc"] Oct 08 16:36:02 crc kubenswrapper[4945]: I1008 16:36:02.348412 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerStarted","Data":"c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151"} Oct 08 16:36:02 crc kubenswrapper[4945]: I1008 16:36:02.348761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerStarted","Data":"8126670a1a4b6e2052e9c075b6fd51d86adcb020f4d09d23dfe2ace2514c4a38"} Oct 08 16:36:03 crc kubenswrapper[4945]: I1008 16:36:03.371598 4945 generic.go:334] "Generic (PLEG): container finished" podID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerID="c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151" exitCode=0 Oct 08 16:36:03 crc kubenswrapper[4945]: I1008 16:36:03.371648 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerDied","Data":"c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151"} Oct 08 16:36:03 crc kubenswrapper[4945]: I1008 16:36:03.372101 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerStarted","Data":"9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a"} Oct 08 16:36:05 crc kubenswrapper[4945]: I1008 16:36:05.393022 4945 generic.go:334] "Generic (PLEG): container finished" podID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerID="9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a" exitCode=0 Oct 08 16:36:05 crc kubenswrapper[4945]: I1008 16:36:05.393122 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerDied","Data":"9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a"} Oct 08 16:36:06 crc kubenswrapper[4945]: I1008 16:36:06.279731 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hmqtn" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="registry-server" probeResult="failure" output=< Oct 08 16:36:06 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:36:06 crc kubenswrapper[4945]: > Oct 08 16:36:06 crc kubenswrapper[4945]: I1008 16:36:06.404815 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerStarted","Data":"2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35"} Oct 08 16:36:06 crc kubenswrapper[4945]: I1008 16:36:06.432383 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mrhnc" podStartSLOduration=1.9365434179999998 podStartE2EDuration="5.432356313s" podCreationTimestamp="2025-10-08 16:36:01 +0000 UTC" firstStartedPulling="2025-10-08 16:36:02.350764498 +0000 UTC m=+4971.704679399" lastFinishedPulling="2025-10-08 16:36:05.846577393 +0000 UTC m=+4975.200492294" observedRunningTime="2025-10-08 16:36:06.423208306 +0000 UTC m=+4975.777123227" watchObservedRunningTime="2025-10-08 16:36:06.432356313 +0000 UTC m=+4975.786271234" Oct 08 16:36:11 crc kubenswrapper[4945]: I1008 16:36:11.539363 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:11 crc kubenswrapper[4945]: I1008 16:36:11.539909 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:12 crc kubenswrapper[4945]: I1008 16:36:12.600666 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mrhnc" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="registry-server" probeResult="failure" output=< Oct 08 16:36:12 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:36:12 crc kubenswrapper[4945]: > Oct 08 16:36:15 crc kubenswrapper[4945]: I1008 16:36:15.301350 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:36:15 crc kubenswrapper[4945]: I1008 16:36:15.350147 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:36:15 crc kubenswrapper[4945]: I1008 16:36:15.537941 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hmqtn"] Oct 08 16:36:16 crc kubenswrapper[4945]: I1008 16:36:16.495077 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hmqtn" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="registry-server" containerID="cri-o://f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067" gracePeriod=2 Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.043413 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.187840 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-catalog-content\") pod \"cb793be2-9d58-4a3b-814f-17efa4392c96\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.188039 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t4wn\" (UniqueName: \"kubernetes.io/projected/cb793be2-9d58-4a3b-814f-17efa4392c96-kube-api-access-2t4wn\") pod \"cb793be2-9d58-4a3b-814f-17efa4392c96\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.188206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-utilities\") pod \"cb793be2-9d58-4a3b-814f-17efa4392c96\" (UID: \"cb793be2-9d58-4a3b-814f-17efa4392c96\") " Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.189098 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-utilities" (OuterVolumeSpecName: "utilities") pod "cb793be2-9d58-4a3b-814f-17efa4392c96" (UID: "cb793be2-9d58-4a3b-814f-17efa4392c96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.197631 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb793be2-9d58-4a3b-814f-17efa4392c96-kube-api-access-2t4wn" (OuterVolumeSpecName: "kube-api-access-2t4wn") pod "cb793be2-9d58-4a3b-814f-17efa4392c96" (UID: "cb793be2-9d58-4a3b-814f-17efa4392c96"). InnerVolumeSpecName "kube-api-access-2t4wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.279713 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb793be2-9d58-4a3b-814f-17efa4392c96" (UID: "cb793be2-9d58-4a3b-814f-17efa4392c96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.290788 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t4wn\" (UniqueName: \"kubernetes.io/projected/cb793be2-9d58-4a3b-814f-17efa4392c96-kube-api-access-2t4wn\") on node \"crc\" DevicePath \"\"" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.290834 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.290847 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb793be2-9d58-4a3b-814f-17efa4392c96-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.506560 4945 generic.go:334] "Generic (PLEG): container finished" podID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerID="f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067" exitCode=0 Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.506682 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmqtn" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.506676 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmqtn" event={"ID":"cb793be2-9d58-4a3b-814f-17efa4392c96","Type":"ContainerDied","Data":"f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067"} Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.507005 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmqtn" event={"ID":"cb793be2-9d58-4a3b-814f-17efa4392c96","Type":"ContainerDied","Data":"b2692a3254d508e8dd2f37222237c7f4c2af1a4db181bad1189a7a5d08eca36c"} Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.507033 4945 scope.go:117] "RemoveContainer" containerID="f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.531261 4945 scope.go:117] "RemoveContainer" containerID="722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.542695 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hmqtn"] Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.551515 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hmqtn"] Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.569420 4945 scope.go:117] "RemoveContainer" containerID="452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.605282 4945 scope.go:117] "RemoveContainer" containerID="f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067" Oct 08 16:36:17 crc kubenswrapper[4945]: E1008 16:36:17.605654 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067\": container with ID starting with f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067 not found: ID does not exist" containerID="f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.605686 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067"} err="failed to get container status \"f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067\": rpc error: code = NotFound desc = could not find container \"f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067\": container with ID starting with f64b149b2a9f4af5ec6683a7e44c7a82b63fbdcf7eb8c8950881dc8d1d88c067 not found: ID does not exist" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.605708 4945 scope.go:117] "RemoveContainer" containerID="722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f" Oct 08 16:36:17 crc kubenswrapper[4945]: E1008 16:36:17.606076 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f\": container with ID starting with 722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f not found: ID does not exist" containerID="722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.606128 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f"} err="failed to get container status \"722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f\": rpc error: code = NotFound desc = could not find container \"722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f\": container with ID starting with 722bd8ad08e639408985a7c6927219745fb4cc8ac6853b88cbfada1cc5abb97f not found: ID does not exist" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.606145 4945 scope.go:117] "RemoveContainer" containerID="452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09" Oct 08 16:36:17 crc kubenswrapper[4945]: E1008 16:36:17.606525 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09\": container with ID starting with 452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09 not found: ID does not exist" containerID="452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09" Oct 08 16:36:17 crc kubenswrapper[4945]: I1008 16:36:17.606551 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09"} err="failed to get container status \"452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09\": rpc error: code = NotFound desc = could not find container \"452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09\": container with ID starting with 452824370b69de981e40ebd841e25d5ef4b60fdd0af8f8d0d486c8721af13c09 not found: ID does not exist" Oct 08 16:36:18 crc kubenswrapper[4945]: I1008 16:36:18.037950 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" path="/var/lib/kubelet/pods/cb793be2-9d58-4a3b-814f-17efa4392c96/volumes" Oct 08 16:36:19 crc kubenswrapper[4945]: I1008 16:36:19.184463 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:36:19 crc kubenswrapper[4945]: I1008 16:36:19.184802 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:36:21 crc kubenswrapper[4945]: I1008 16:36:21.585847 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:21 crc kubenswrapper[4945]: I1008 16:36:21.632786 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:21 crc kubenswrapper[4945]: I1008 16:36:21.824819 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrhnc"] Oct 08 16:36:23 crc kubenswrapper[4945]: I1008 16:36:23.565460 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mrhnc" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="registry-server" containerID="cri-o://2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35" gracePeriod=2 Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.083000 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.231917 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-utilities\") pod \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.232396 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6pzj\" (UniqueName: \"kubernetes.io/projected/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-kube-api-access-d6pzj\") pod \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.232436 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-catalog-content\") pod \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\" (UID: \"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c\") " Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.232959 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-utilities" (OuterVolumeSpecName: "utilities") pod "ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" (UID: "ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.233170 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.242333 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-kube-api-access-d6pzj" (OuterVolumeSpecName: "kube-api-access-d6pzj") pod "ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" (UID: "ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c"). InnerVolumeSpecName "kube-api-access-d6pzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.278868 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" (UID: "ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.334781 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6pzj\" (UniqueName: \"kubernetes.io/projected/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-kube-api-access-d6pzj\") on node \"crc\" DevicePath \"\"" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.334816 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.578003 4945 generic.go:334] "Generic (PLEG): container finished" podID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerID="2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35" exitCode=0 Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.578044 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerDied","Data":"2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35"} Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.578074 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhnc" event={"ID":"ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c","Type":"ContainerDied","Data":"8126670a1a4b6e2052e9c075b6fd51d86adcb020f4d09d23dfe2ace2514c4a38"} Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.578092 4945 scope.go:117] "RemoveContainer" containerID="2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.578157 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhnc" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.602629 4945 scope.go:117] "RemoveContainer" containerID="9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.616124 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrhnc"] Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.623745 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mrhnc"] Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.649659 4945 scope.go:117] "RemoveContainer" containerID="c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.688089 4945 scope.go:117] "RemoveContainer" containerID="2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35" Oct 08 16:36:24 crc kubenswrapper[4945]: E1008 16:36:24.688426 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35\": container with ID starting with 2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35 not found: ID does not exist" containerID="2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.688478 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35"} err="failed to get container status \"2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35\": rpc error: code = NotFound desc = could not find container \"2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35\": container with ID starting with 2cde7939c278d608d859a17395c4d1ba1bcdd8d17fb33c0832770f9d1a719b35 not found: ID does not exist" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.688506 4945 scope.go:117] "RemoveContainer" containerID="9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a" Oct 08 16:36:24 crc kubenswrapper[4945]: E1008 16:36:24.688804 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a\": container with ID starting with 9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a not found: ID does not exist" containerID="9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.688840 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a"} err="failed to get container status \"9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a\": rpc error: code = NotFound desc = could not find container \"9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a\": container with ID starting with 9e6d6f1e16f28daf4a5dfa438eaffe027b987007cedcf23421739b4f815f715a not found: ID does not exist" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.688872 4945 scope.go:117] "RemoveContainer" containerID="c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151" Oct 08 16:36:24 crc kubenswrapper[4945]: E1008 16:36:24.689310 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151\": container with ID starting with c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151 not found: ID does not exist" containerID="c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151" Oct 08 16:36:24 crc kubenswrapper[4945]: I1008 16:36:24.689342 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151"} err="failed to get container status \"c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151\": rpc error: code = NotFound desc = could not find container \"c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151\": container with ID starting with c0e42f5168fd19ac5f9fda6837148138834aeb3a8b58f66235c6aee6f452c151 not found: ID does not exist" Oct 08 16:36:26 crc kubenswrapper[4945]: I1008 16:36:26.037083 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" path="/var/lib/kubelet/pods/ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c/volumes" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.184521 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.185085 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.640267 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j4btv"] Oct 08 16:36:49 crc kubenswrapper[4945]: E1008 16:36:49.640770 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="extract-utilities" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.640795 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="extract-utilities" Oct 08 16:36:49 crc kubenswrapper[4945]: E1008 16:36:49.640813 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="registry-server" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.640821 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="registry-server" Oct 08 16:36:49 crc kubenswrapper[4945]: E1008 16:36:49.640846 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="registry-server" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.640855 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="registry-server" Oct 08 16:36:49 crc kubenswrapper[4945]: E1008 16:36:49.640873 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="extract-content" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.640880 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="extract-content" Oct 08 16:36:49 crc kubenswrapper[4945]: E1008 16:36:49.640894 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="extract-content" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.640901 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="extract-content" Oct 08 16:36:49 crc kubenswrapper[4945]: E1008 16:36:49.640921 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="extract-utilities" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.640930 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="extract-utilities" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.641162 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad37aef9-ce4f-47f9-9b62-7f5f4cb2c81c" containerName="registry-server" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.641197 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb793be2-9d58-4a3b-814f-17efa4392c96" containerName="registry-server" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.643005 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.653950 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4btv"] Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.756543 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-catalog-content\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.756643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-utilities\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.756713 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtm52\" (UniqueName: \"kubernetes.io/projected/38c601f3-72af-497e-8a6d-21739c661c8c-kube-api-access-vtm52\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.859244 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtm52\" (UniqueName: \"kubernetes.io/projected/38c601f3-72af-497e-8a6d-21739c661c8c-kube-api-access-vtm52\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.859452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-catalog-content\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.859519 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-utilities\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.860067 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-catalog-content\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.860155 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-utilities\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.877848 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtm52\" (UniqueName: \"kubernetes.io/projected/38c601f3-72af-497e-8a6d-21739c661c8c-kube-api-access-vtm52\") pod \"redhat-marketplace-j4btv\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:49 crc kubenswrapper[4945]: I1008 16:36:49.978353 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:50 crc kubenswrapper[4945]: I1008 16:36:50.448817 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4btv"] Oct 08 16:36:50 crc kubenswrapper[4945]: I1008 16:36:50.819791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4btv" event={"ID":"38c601f3-72af-497e-8a6d-21739c661c8c","Type":"ContainerStarted","Data":"e12eca43ebbd0cee11e68a13818faafcf7f08a44f3c739de9e1355fd2f97f860"} Oct 08 16:36:51 crc kubenswrapper[4945]: I1008 16:36:51.831998 4945 generic.go:334] "Generic (PLEG): container finished" podID="38c601f3-72af-497e-8a6d-21739c661c8c" containerID="f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c" exitCode=0 Oct 08 16:36:51 crc kubenswrapper[4945]: I1008 16:36:51.832036 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4btv" event={"ID":"38c601f3-72af-497e-8a6d-21739c661c8c","Type":"ContainerDied","Data":"f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c"} Oct 08 16:36:53 crc kubenswrapper[4945]: I1008 16:36:53.853453 4945 generic.go:334] "Generic (PLEG): container finished" podID="38c601f3-72af-497e-8a6d-21739c661c8c" containerID="2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094" exitCode=0 Oct 08 16:36:53 crc kubenswrapper[4945]: I1008 16:36:53.853519 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4btv" event={"ID":"38c601f3-72af-497e-8a6d-21739c661c8c","Type":"ContainerDied","Data":"2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094"} Oct 08 16:36:54 crc kubenswrapper[4945]: I1008 16:36:54.867176 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4btv" event={"ID":"38c601f3-72af-497e-8a6d-21739c661c8c","Type":"ContainerStarted","Data":"d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9"} Oct 08 16:36:54 crc kubenswrapper[4945]: I1008 16:36:54.890996 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j4btv" podStartSLOduration=3.216445137 podStartE2EDuration="5.890977178s" podCreationTimestamp="2025-10-08 16:36:49 +0000 UTC" firstStartedPulling="2025-10-08 16:36:51.837155736 +0000 UTC m=+5021.191070667" lastFinishedPulling="2025-10-08 16:36:54.511687807 +0000 UTC m=+5023.865602708" observedRunningTime="2025-10-08 16:36:54.883724138 +0000 UTC m=+5024.237639049" watchObservedRunningTime="2025-10-08 16:36:54.890977178 +0000 UTC m=+5024.244892079" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.310076 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-828sq"] Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.319497 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.344299 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-828sq"] Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.410890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-utilities\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.411359 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-catalog-content\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.411629 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m94n6\" (UniqueName: \"kubernetes.io/projected/f0f73aba-75f6-406f-91f2-7a8c2f350eba-kube-api-access-m94n6\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.513369 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-utilities\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.513546 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-catalog-content\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.513683 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m94n6\" (UniqueName: \"kubernetes.io/projected/f0f73aba-75f6-406f-91f2-7a8c2f350eba-kube-api-access-m94n6\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.513990 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-utilities\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.514081 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-catalog-content\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.535153 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m94n6\" (UniqueName: \"kubernetes.io/projected/f0f73aba-75f6-406f-91f2-7a8c2f350eba-kube-api-access-m94n6\") pod \"community-operators-828sq\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:57 crc kubenswrapper[4945]: I1008 16:36:57.644625 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:36:58 crc kubenswrapper[4945]: I1008 16:36:58.202357 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-828sq"] Oct 08 16:36:58 crc kubenswrapper[4945]: I1008 16:36:58.909427 4945 generic.go:334] "Generic (PLEG): container finished" podID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerID="1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88" exitCode=0 Oct 08 16:36:58 crc kubenswrapper[4945]: I1008 16:36:58.909482 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-828sq" event={"ID":"f0f73aba-75f6-406f-91f2-7a8c2f350eba","Type":"ContainerDied","Data":"1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88"} Oct 08 16:36:58 crc kubenswrapper[4945]: I1008 16:36:58.909739 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-828sq" event={"ID":"f0f73aba-75f6-406f-91f2-7a8c2f350eba","Type":"ContainerStarted","Data":"d84c9408f95e9561cc31e8d9541384a3553a38f37c9430e8fe879af85f2772b4"} Oct 08 16:36:59 crc kubenswrapper[4945]: I1008 16:36:59.979360 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:36:59 crc kubenswrapper[4945]: I1008 16:36:59.979832 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:37:00 crc kubenswrapper[4945]: I1008 16:37:00.040501 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:37:00 crc kubenswrapper[4945]: I1008 16:37:00.950649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-828sq" event={"ID":"f0f73aba-75f6-406f-91f2-7a8c2f350eba","Type":"ContainerStarted","Data":"12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635"} Oct 08 16:37:01 crc kubenswrapper[4945]: I1008 16:37:01.020302 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:37:01 crc kubenswrapper[4945]: E1008 16:37:01.124374 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0f73aba_75f6_406f_91f2_7a8c2f350eba.slice/crio-conmon-12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635.scope\": RecentStats: unable to find data in memory cache]" Oct 08 16:37:01 crc kubenswrapper[4945]: I1008 16:37:01.964754 4945 generic.go:334] "Generic (PLEG): container finished" podID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerID="12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635" exitCode=0 Oct 08 16:37:01 crc kubenswrapper[4945]: I1008 16:37:01.964798 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-828sq" event={"ID":"f0f73aba-75f6-406f-91f2-7a8c2f350eba","Type":"ContainerDied","Data":"12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635"} Oct 08 16:37:02 crc kubenswrapper[4945]: I1008 16:37:02.282051 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4btv"] Oct 08 16:37:02 crc kubenswrapper[4945]: I1008 16:37:02.976550 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j4btv" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="registry-server" containerID="cri-o://d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9" gracePeriod=2 Oct 08 16:37:02 crc kubenswrapper[4945]: I1008 16:37:02.976592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-828sq" event={"ID":"f0f73aba-75f6-406f-91f2-7a8c2f350eba","Type":"ContainerStarted","Data":"002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13"} Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.009148 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-828sq" podStartSLOduration=2.468531938 podStartE2EDuration="6.009128178s" podCreationTimestamp="2025-10-08 16:36:57 +0000 UTC" firstStartedPulling="2025-10-08 16:36:58.915653946 +0000 UTC m=+5028.269568857" lastFinishedPulling="2025-10-08 16:37:02.456250196 +0000 UTC m=+5031.810165097" observedRunningTime="2025-10-08 16:37:03.000494503 +0000 UTC m=+5032.354409404" watchObservedRunningTime="2025-10-08 16:37:03.009128178 +0000 UTC m=+5032.363043079" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.500212 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.586578 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-catalog-content\") pod \"38c601f3-72af-497e-8a6d-21739c661c8c\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.586684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-utilities\") pod \"38c601f3-72af-497e-8a6d-21739c661c8c\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.586732 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtm52\" (UniqueName: \"kubernetes.io/projected/38c601f3-72af-497e-8a6d-21739c661c8c-kube-api-access-vtm52\") pod \"38c601f3-72af-497e-8a6d-21739c661c8c\" (UID: \"38c601f3-72af-497e-8a6d-21739c661c8c\") " Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.587730 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-utilities" (OuterVolumeSpecName: "utilities") pod "38c601f3-72af-497e-8a6d-21739c661c8c" (UID: "38c601f3-72af-497e-8a6d-21739c661c8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.594026 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c601f3-72af-497e-8a6d-21739c661c8c-kube-api-access-vtm52" (OuterVolumeSpecName: "kube-api-access-vtm52") pod "38c601f3-72af-497e-8a6d-21739c661c8c" (UID: "38c601f3-72af-497e-8a6d-21739c661c8c"). InnerVolumeSpecName "kube-api-access-vtm52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.604018 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38c601f3-72af-497e-8a6d-21739c661c8c" (UID: "38c601f3-72af-497e-8a6d-21739c661c8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.689707 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.689750 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c601f3-72af-497e-8a6d-21739c661c8c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.689760 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtm52\" (UniqueName: \"kubernetes.io/projected/38c601f3-72af-497e-8a6d-21739c661c8c-kube-api-access-vtm52\") on node \"crc\" DevicePath \"\"" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.989655 4945 generic.go:334] "Generic (PLEG): container finished" podID="38c601f3-72af-497e-8a6d-21739c661c8c" containerID="d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9" exitCode=0 Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.989712 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4btv" Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.989734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4btv" event={"ID":"38c601f3-72af-497e-8a6d-21739c661c8c","Type":"ContainerDied","Data":"d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9"} Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.990414 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4btv" event={"ID":"38c601f3-72af-497e-8a6d-21739c661c8c","Type":"ContainerDied","Data":"e12eca43ebbd0cee11e68a13818faafcf7f08a44f3c739de9e1355fd2f97f860"} Oct 08 16:37:03 crc kubenswrapper[4945]: I1008 16:37:03.990484 4945 scope.go:117] "RemoveContainer" containerID="d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.042981 4945 scope.go:117] "RemoveContainer" containerID="2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.047833 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4btv"] Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.057519 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4btv"] Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.076227 4945 scope.go:117] "RemoveContainer" containerID="f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.130787 4945 scope.go:117] "RemoveContainer" containerID="d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9" Oct 08 16:37:04 crc kubenswrapper[4945]: E1008 16:37:04.131253 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9\": container with ID starting with d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9 not found: ID does not exist" containerID="d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.131301 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9"} err="failed to get container status \"d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9\": rpc error: code = NotFound desc = could not find container \"d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9\": container with ID starting with d854eae0d2ce9dec2823b7d4a08eaf27839a2ecc99ebf7abf000dcf987012eb9 not found: ID does not exist" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.131328 4945 scope.go:117] "RemoveContainer" containerID="2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094" Oct 08 16:37:04 crc kubenswrapper[4945]: E1008 16:37:04.131577 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094\": container with ID starting with 2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094 not found: ID does not exist" containerID="2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.131602 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094"} err="failed to get container status \"2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094\": rpc error: code = NotFound desc = could not find container \"2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094\": container with ID starting with 2398d41552147d376fcab801c4afa944c33ebadaa35328a0fe8d2090e4265094 not found: ID does not exist" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.131617 4945 scope.go:117] "RemoveContainer" containerID="f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c" Oct 08 16:37:04 crc kubenswrapper[4945]: E1008 16:37:04.131815 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c\": container with ID starting with f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c not found: ID does not exist" containerID="f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c" Oct 08 16:37:04 crc kubenswrapper[4945]: I1008 16:37:04.131835 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c"} err="failed to get container status \"f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c\": rpc error: code = NotFound desc = could not find container \"f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c\": container with ID starting with f986ec3cd36691c9f15f17ba7d3730d749f05e1cdf64d646f7c2d087c707316c not found: ID does not exist" Oct 08 16:37:06 crc kubenswrapper[4945]: I1008 16:37:06.046058 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" path="/var/lib/kubelet/pods/38c601f3-72af-497e-8a6d-21739c661c8c/volumes" Oct 08 16:37:07 crc kubenswrapper[4945]: I1008 16:37:07.645770 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:37:07 crc kubenswrapper[4945]: I1008 16:37:07.645824 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:37:07 crc kubenswrapper[4945]: I1008 16:37:07.697456 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:37:08 crc kubenswrapper[4945]: I1008 16:37:08.090209 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:37:08 crc kubenswrapper[4945]: I1008 16:37:08.880335 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-828sq"] Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.056226 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-828sq" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="registry-server" containerID="cri-o://002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13" gracePeriod=2 Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.545765 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.637331 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-utilities\") pod \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.637456 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-catalog-content\") pod \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.637484 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m94n6\" (UniqueName: \"kubernetes.io/projected/f0f73aba-75f6-406f-91f2-7a8c2f350eba-kube-api-access-m94n6\") pod \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\" (UID: \"f0f73aba-75f6-406f-91f2-7a8c2f350eba\") " Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.638074 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-utilities" (OuterVolumeSpecName: "utilities") pod "f0f73aba-75f6-406f-91f2-7a8c2f350eba" (UID: "f0f73aba-75f6-406f-91f2-7a8c2f350eba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.643237 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0f73aba-75f6-406f-91f2-7a8c2f350eba-kube-api-access-m94n6" (OuterVolumeSpecName: "kube-api-access-m94n6") pod "f0f73aba-75f6-406f-91f2-7a8c2f350eba" (UID: "f0f73aba-75f6-406f-91f2-7a8c2f350eba"). InnerVolumeSpecName "kube-api-access-m94n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.691223 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0f73aba-75f6-406f-91f2-7a8c2f350eba" (UID: "f0f73aba-75f6-406f-91f2-7a8c2f350eba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.739745 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.739779 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0f73aba-75f6-406f-91f2-7a8c2f350eba-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:37:10 crc kubenswrapper[4945]: I1008 16:37:10.739790 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m94n6\" (UniqueName: \"kubernetes.io/projected/f0f73aba-75f6-406f-91f2-7a8c2f350eba-kube-api-access-m94n6\") on node \"crc\" DevicePath \"\"" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.066579 4945 generic.go:334] "Generic (PLEG): container finished" podID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerID="002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13" exitCode=0 Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.066646 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-828sq" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.066642 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-828sq" event={"ID":"f0f73aba-75f6-406f-91f2-7a8c2f350eba","Type":"ContainerDied","Data":"002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13"} Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.066703 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-828sq" event={"ID":"f0f73aba-75f6-406f-91f2-7a8c2f350eba","Type":"ContainerDied","Data":"d84c9408f95e9561cc31e8d9541384a3553a38f37c9430e8fe879af85f2772b4"} Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.066726 4945 scope.go:117] "RemoveContainer" containerID="002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.108100 4945 scope.go:117] "RemoveContainer" containerID="12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.119258 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-828sq"] Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.136449 4945 scope.go:117] "RemoveContainer" containerID="1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.138205 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-828sq"] Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.187914 4945 scope.go:117] "RemoveContainer" containerID="002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13" Oct 08 16:37:11 crc kubenswrapper[4945]: E1008 16:37:11.188647 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13\": container with ID starting with 002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13 not found: ID does not exist" containerID="002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.188689 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13"} err="failed to get container status \"002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13\": rpc error: code = NotFound desc = could not find container \"002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13\": container with ID starting with 002cd61f0c2021f5bb4ceaa8a30b102952bfd8f9320425746dfad4ff33751c13 not found: ID does not exist" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.188716 4945 scope.go:117] "RemoveContainer" containerID="12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635" Oct 08 16:37:11 crc kubenswrapper[4945]: E1008 16:37:11.189067 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635\": container with ID starting with 12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635 not found: ID does not exist" containerID="12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.189094 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635"} err="failed to get container status \"12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635\": rpc error: code = NotFound desc = could not find container \"12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635\": container with ID starting with 12a7fd48a1287426ab7644255460cf7f4a823a82a814d6407cc5b686e8c23635 not found: ID does not exist" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.189127 4945 scope.go:117] "RemoveContainer" containerID="1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88" Oct 08 16:37:11 crc kubenswrapper[4945]: E1008 16:37:11.189538 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88\": container with ID starting with 1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88 not found: ID does not exist" containerID="1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88" Oct 08 16:37:11 crc kubenswrapper[4945]: I1008 16:37:11.189565 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88"} err="failed to get container status \"1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88\": rpc error: code = NotFound desc = could not find container \"1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88\": container with ID starting with 1840b6a72d319c43d896797e8e65b33587dca0c743bfff92165a5c6bf8b07c88 not found: ID does not exist" Oct 08 16:37:12 crc kubenswrapper[4945]: I1008 16:37:12.041305 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" path="/var/lib/kubelet/pods/f0f73aba-75f6-406f-91f2-7a8c2f350eba/volumes" Oct 08 16:37:19 crc kubenswrapper[4945]: I1008 16:37:19.184867 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:37:19 crc kubenswrapper[4945]: I1008 16:37:19.185439 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:37:19 crc kubenswrapper[4945]: I1008 16:37:19.185483 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:37:19 crc kubenswrapper[4945]: I1008 16:37:19.186254 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:37:19 crc kubenswrapper[4945]: I1008 16:37:19.186311 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" gracePeriod=600 Oct 08 16:37:19 crc kubenswrapper[4945]: E1008 16:37:19.309334 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:37:20 crc kubenswrapper[4945]: I1008 16:37:20.158812 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" exitCode=0 Oct 08 16:37:20 crc kubenswrapper[4945]: I1008 16:37:20.158854 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4"} Oct 08 16:37:20 crc kubenswrapper[4945]: I1008 16:37:20.158888 4945 scope.go:117] "RemoveContainer" containerID="2a3055c35e39f4c78367daaba70caaada0e749323c9fd679bc600a59fa8a8233" Oct 08 16:37:20 crc kubenswrapper[4945]: I1008 16:37:20.159663 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:37:20 crc kubenswrapper[4945]: E1008 16:37:20.159918 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:37:35 crc kubenswrapper[4945]: I1008 16:37:35.024537 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:37:35 crc kubenswrapper[4945]: E1008 16:37:35.025335 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:37:48 crc kubenswrapper[4945]: I1008 16:37:48.024831 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:37:48 crc kubenswrapper[4945]: E1008 16:37:48.025758 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:38:01 crc kubenswrapper[4945]: I1008 16:38:01.024770 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:38:01 crc kubenswrapper[4945]: E1008 16:38:01.025593 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:38:13 crc kubenswrapper[4945]: I1008 16:38:13.024818 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:38:13 crc kubenswrapper[4945]: E1008 16:38:13.027771 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:38:24 crc kubenswrapper[4945]: I1008 16:38:24.026367 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:38:24 crc kubenswrapper[4945]: E1008 16:38:24.027139 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:38:38 crc kubenswrapper[4945]: I1008 16:38:38.024971 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:38:38 crc kubenswrapper[4945]: E1008 16:38:38.025849 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:38:52 crc kubenswrapper[4945]: I1008 16:38:52.040712 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:38:52 crc kubenswrapper[4945]: E1008 16:38:52.042053 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:39:06 crc kubenswrapper[4945]: I1008 16:39:06.024530 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:39:06 crc kubenswrapper[4945]: E1008 16:39:06.025478 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:39:18 crc kubenswrapper[4945]: I1008 16:39:18.024195 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:39:18 crc kubenswrapper[4945]: E1008 16:39:18.025085 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:39:30 crc kubenswrapper[4945]: I1008 16:39:30.024192 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:39:30 crc kubenswrapper[4945]: E1008 16:39:30.025136 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:39:44 crc kubenswrapper[4945]: I1008 16:39:44.025085 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:39:44 crc kubenswrapper[4945]: E1008 16:39:44.025998 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:39:58 crc kubenswrapper[4945]: I1008 16:39:58.024341 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:39:58 crc kubenswrapper[4945]: E1008 16:39:58.025178 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:40:10 crc kubenswrapper[4945]: I1008 16:40:10.024609 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:40:10 crc kubenswrapper[4945]: E1008 16:40:10.025621 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:40:22 crc kubenswrapper[4945]: I1008 16:40:22.034288 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:40:22 crc kubenswrapper[4945]: E1008 16:40:22.034950 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:40:33 crc kubenswrapper[4945]: I1008 16:40:33.024324 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:40:33 crc kubenswrapper[4945]: E1008 16:40:33.025454 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:40:48 crc kubenswrapper[4945]: I1008 16:40:48.026342 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:40:48 crc kubenswrapper[4945]: E1008 16:40:48.037132 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:41:03 crc kubenswrapper[4945]: I1008 16:41:03.024430 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:41:03 crc kubenswrapper[4945]: E1008 16:41:03.025312 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:41:17 crc kubenswrapper[4945]: I1008 16:41:17.025224 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:41:17 crc kubenswrapper[4945]: E1008 16:41:17.026164 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:41:31 crc kubenswrapper[4945]: I1008 16:41:31.024053 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:41:31 crc kubenswrapper[4945]: E1008 16:41:31.024847 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:41:45 crc kubenswrapper[4945]: I1008 16:41:45.026141 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:41:45 crc kubenswrapper[4945]: E1008 16:41:45.027427 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:41:58 crc kubenswrapper[4945]: I1008 16:41:58.025577 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:41:58 crc kubenswrapper[4945]: E1008 16:41:58.027060 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:42:13 crc kubenswrapper[4945]: I1008 16:42:13.025074 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:42:13 crc kubenswrapper[4945]: E1008 16:42:13.026375 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:42:26 crc kubenswrapper[4945]: I1008 16:42:26.024629 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:42:27 crc kubenswrapper[4945]: I1008 16:42:27.247611 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"a9568695c8a4dca15d9f95043c6fb8f3e6a9782b4a613574a9f0241ad503097e"} Oct 08 16:44:49 crc kubenswrapper[4945]: I1008 16:44:49.184706 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:44:49 crc kubenswrapper[4945]: I1008 16:44:49.185342 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.154980 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm"] Oct 08 16:45:00 crc kubenswrapper[4945]: E1008 16:45:00.156262 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="extract-utilities" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156281 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="extract-utilities" Oct 08 16:45:00 crc kubenswrapper[4945]: E1008 16:45:00.156320 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="extract-utilities" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156328 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="extract-utilities" Oct 08 16:45:00 crc kubenswrapper[4945]: E1008 16:45:00.156344 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="registry-server" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156352 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="registry-server" Oct 08 16:45:00 crc kubenswrapper[4945]: E1008 16:45:00.156367 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="extract-content" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156375 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="extract-content" Oct 08 16:45:00 crc kubenswrapper[4945]: E1008 16:45:00.156390 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="extract-content" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156397 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="extract-content" Oct 08 16:45:00 crc kubenswrapper[4945]: E1008 16:45:00.156426 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="registry-server" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156434 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="registry-server" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156715 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c601f3-72af-497e-8a6d-21739c661c8c" containerName="registry-server" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.156740 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0f73aba-75f6-406f-91f2-7a8c2f350eba" containerName="registry-server" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.157689 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.160261 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.160664 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.166851 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm"] Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.231557 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d9195f3-aea9-468b-8072-6691f6739514-config-volume\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.231694 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d9195f3-aea9-468b-8072-6691f6739514-secret-volume\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.231734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69t2k\" (UniqueName: \"kubernetes.io/projected/4d9195f3-aea9-468b-8072-6691f6739514-kube-api-access-69t2k\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.333862 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d9195f3-aea9-468b-8072-6691f6739514-secret-volume\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.334238 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69t2k\" (UniqueName: \"kubernetes.io/projected/4d9195f3-aea9-468b-8072-6691f6739514-kube-api-access-69t2k\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.334591 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d9195f3-aea9-468b-8072-6691f6739514-config-volume\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.335522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d9195f3-aea9-468b-8072-6691f6739514-config-volume\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.349357 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d9195f3-aea9-468b-8072-6691f6739514-secret-volume\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.365326 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69t2k\" (UniqueName: \"kubernetes.io/projected/4d9195f3-aea9-468b-8072-6691f6739514-kube-api-access-69t2k\") pod \"collect-profiles-29332365-r6vmm\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.481708 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:00 crc kubenswrapper[4945]: I1008 16:45:00.950343 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm"] Oct 08 16:45:00 crc kubenswrapper[4945]: W1008 16:45:00.968026 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d9195f3_aea9_468b_8072_6691f6739514.slice/crio-ea67487fcef55d5e19ed52689fc6b514677d86548d0ad455c543bd16d9edd0ec WatchSource:0}: Error finding container ea67487fcef55d5e19ed52689fc6b514677d86548d0ad455c543bd16d9edd0ec: Status 404 returned error can't find the container with id ea67487fcef55d5e19ed52689fc6b514677d86548d0ad455c543bd16d9edd0ec Oct 08 16:45:01 crc kubenswrapper[4945]: I1008 16:45:01.748699 4945 generic.go:334] "Generic (PLEG): container finished" podID="4d9195f3-aea9-468b-8072-6691f6739514" containerID="2baf1719a4d00f60774d6356e427039952350dfdf20af6a27c85a14dd1ee1048" exitCode=0 Oct 08 16:45:01 crc kubenswrapper[4945]: I1008 16:45:01.749258 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" event={"ID":"4d9195f3-aea9-468b-8072-6691f6739514","Type":"ContainerDied","Data":"2baf1719a4d00f60774d6356e427039952350dfdf20af6a27c85a14dd1ee1048"} Oct 08 16:45:01 crc kubenswrapper[4945]: I1008 16:45:01.749283 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" event={"ID":"4d9195f3-aea9-468b-8072-6691f6739514","Type":"ContainerStarted","Data":"ea67487fcef55d5e19ed52689fc6b514677d86548d0ad455c543bd16d9edd0ec"} Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.229847 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.296951 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69t2k\" (UniqueName: \"kubernetes.io/projected/4d9195f3-aea9-468b-8072-6691f6739514-kube-api-access-69t2k\") pod \"4d9195f3-aea9-468b-8072-6691f6739514\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.297035 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d9195f3-aea9-468b-8072-6691f6739514-config-volume\") pod \"4d9195f3-aea9-468b-8072-6691f6739514\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.297162 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d9195f3-aea9-468b-8072-6691f6739514-secret-volume\") pod \"4d9195f3-aea9-468b-8072-6691f6739514\" (UID: \"4d9195f3-aea9-468b-8072-6691f6739514\") " Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.298055 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d9195f3-aea9-468b-8072-6691f6739514-config-volume" (OuterVolumeSpecName: "config-volume") pod "4d9195f3-aea9-468b-8072-6691f6739514" (UID: "4d9195f3-aea9-468b-8072-6691f6739514"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.304208 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9195f3-aea9-468b-8072-6691f6739514-kube-api-access-69t2k" (OuterVolumeSpecName: "kube-api-access-69t2k") pod "4d9195f3-aea9-468b-8072-6691f6739514" (UID: "4d9195f3-aea9-468b-8072-6691f6739514"). InnerVolumeSpecName "kube-api-access-69t2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.307083 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9195f3-aea9-468b-8072-6691f6739514-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4d9195f3-aea9-468b-8072-6691f6739514" (UID: "4d9195f3-aea9-468b-8072-6691f6739514"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.400119 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69t2k\" (UniqueName: \"kubernetes.io/projected/4d9195f3-aea9-468b-8072-6691f6739514-kube-api-access-69t2k\") on node \"crc\" DevicePath \"\"" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.400159 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d9195f3-aea9-468b-8072-6691f6739514-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.400171 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d9195f3-aea9-468b-8072-6691f6739514-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.770941 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" event={"ID":"4d9195f3-aea9-468b-8072-6691f6739514","Type":"ContainerDied","Data":"ea67487fcef55d5e19ed52689fc6b514677d86548d0ad455c543bd16d9edd0ec"} Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.771321 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea67487fcef55d5e19ed52689fc6b514677d86548d0ad455c543bd16d9edd0ec" Oct 08 16:45:03 crc kubenswrapper[4945]: I1008 16:45:03.771036 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332365-r6vmm" Oct 08 16:45:03 crc kubenswrapper[4945]: E1008 16:45:03.983918 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d9195f3_aea9_468b_8072_6691f6739514.slice\": RecentStats: unable to find data in memory cache]" Oct 08 16:45:04 crc kubenswrapper[4945]: I1008 16:45:04.320540 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx"] Oct 08 16:45:04 crc kubenswrapper[4945]: I1008 16:45:04.331143 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332320-rwnqx"] Oct 08 16:45:06 crc kubenswrapper[4945]: I1008 16:45:06.044367 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a696c238-b2c9-4d8b-9977-08ea3c148dbe" path="/var/lib/kubelet/pods/a696c238-b2c9-4d8b-9977-08ea3c148dbe/volumes" Oct 08 16:45:19 crc kubenswrapper[4945]: I1008 16:45:19.183969 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:45:19 crc kubenswrapper[4945]: I1008 16:45:19.184618 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.192976 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hxdzq"] Oct 08 16:45:42 crc kubenswrapper[4945]: E1008 16:45:42.194229 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9195f3-aea9-468b-8072-6691f6739514" containerName="collect-profiles" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.194245 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9195f3-aea9-468b-8072-6691f6739514" containerName="collect-profiles" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.194466 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9195f3-aea9-468b-8072-6691f6739514" containerName="collect-profiles" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.196781 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.204929 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzq"] Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.220410 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-utilities\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.220620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pbxd\" (UniqueName: \"kubernetes.io/projected/03117354-3316-41c5-b997-b24a64f6643e-kube-api-access-5pbxd\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.220793 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-catalog-content\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.321746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-utilities\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.321824 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pbxd\" (UniqueName: \"kubernetes.io/projected/03117354-3316-41c5-b997-b24a64f6643e-kube-api-access-5pbxd\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.321885 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-catalog-content\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.322448 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-catalog-content\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.322508 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-utilities\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.356570 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pbxd\" (UniqueName: \"kubernetes.io/projected/03117354-3316-41c5-b997-b24a64f6643e-kube-api-access-5pbxd\") pod \"redhat-operators-hxdzq\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:42 crc kubenswrapper[4945]: I1008 16:45:42.517562 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:43 crc kubenswrapper[4945]: I1008 16:45:43.047995 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzq"] Oct 08 16:45:43 crc kubenswrapper[4945]: I1008 16:45:43.171195 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzq" event={"ID":"03117354-3316-41c5-b997-b24a64f6643e","Type":"ContainerStarted","Data":"8ae683a4bccc3613100d008cc7bf4d80052324ca8c7540d27e51aba85afa9d05"} Oct 08 16:45:44 crc kubenswrapper[4945]: I1008 16:45:44.180949 4945 generic.go:334] "Generic (PLEG): container finished" podID="03117354-3316-41c5-b997-b24a64f6643e" containerID="5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007" exitCode=0 Oct 08 16:45:44 crc kubenswrapper[4945]: I1008 16:45:44.181206 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzq" event={"ID":"03117354-3316-41c5-b997-b24a64f6643e","Type":"ContainerDied","Data":"5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007"} Oct 08 16:45:44 crc kubenswrapper[4945]: I1008 16:45:44.183654 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 16:45:45 crc kubenswrapper[4945]: I1008 16:45:45.193613 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzq" event={"ID":"03117354-3316-41c5-b997-b24a64f6643e","Type":"ContainerStarted","Data":"78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463"} Oct 08 16:45:47 crc kubenswrapper[4945]: I1008 16:45:47.029044 4945 scope.go:117] "RemoveContainer" containerID="e68e9c13cde3d93ea02f97d78c0e8185e826e68d0087f62081435d2994c4ec19" Oct 08 16:45:49 crc kubenswrapper[4945]: I1008 16:45:49.184075 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:45:49 crc kubenswrapper[4945]: I1008 16:45:49.184657 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:45:49 crc kubenswrapper[4945]: I1008 16:45:49.184709 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:45:49 crc kubenswrapper[4945]: I1008 16:45:49.185527 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9568695c8a4dca15d9f95043c6fb8f3e6a9782b4a613574a9f0241ad503097e"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:45:49 crc kubenswrapper[4945]: I1008 16:45:49.185583 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://a9568695c8a4dca15d9f95043c6fb8f3e6a9782b4a613574a9f0241ad503097e" gracePeriod=600 Oct 08 16:45:49 crc kubenswrapper[4945]: I1008 16:45:49.246651 4945 generic.go:334] "Generic (PLEG): container finished" podID="03117354-3316-41c5-b997-b24a64f6643e" containerID="78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463" exitCode=0 Oct 08 16:45:49 crc kubenswrapper[4945]: I1008 16:45:49.246691 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzq" event={"ID":"03117354-3316-41c5-b997-b24a64f6643e","Type":"ContainerDied","Data":"78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463"} Oct 08 16:45:50 crc kubenswrapper[4945]: I1008 16:45:50.259128 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="a9568695c8a4dca15d9f95043c6fb8f3e6a9782b4a613574a9f0241ad503097e" exitCode=0 Oct 08 16:45:50 crc kubenswrapper[4945]: I1008 16:45:50.259186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"a9568695c8a4dca15d9f95043c6fb8f3e6a9782b4a613574a9f0241ad503097e"} Oct 08 16:45:50 crc kubenswrapper[4945]: I1008 16:45:50.260592 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1"} Oct 08 16:45:50 crc kubenswrapper[4945]: I1008 16:45:50.260622 4945 scope.go:117] "RemoveContainer" containerID="1cf3a58e912fb9b13179a8bd93ad3d3e06141bb7e0b24e9763b77dac8b6673a4" Oct 08 16:45:50 crc kubenswrapper[4945]: I1008 16:45:50.263711 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzq" event={"ID":"03117354-3316-41c5-b997-b24a64f6643e","Type":"ContainerStarted","Data":"f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3"} Oct 08 16:45:50 crc kubenswrapper[4945]: I1008 16:45:50.313914 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hxdzq" podStartSLOduration=2.554348935 podStartE2EDuration="8.31389097s" podCreationTimestamp="2025-10-08 16:45:42 +0000 UTC" firstStartedPulling="2025-10-08 16:45:44.183404405 +0000 UTC m=+5553.537319306" lastFinishedPulling="2025-10-08 16:45:49.94294643 +0000 UTC m=+5559.296861341" observedRunningTime="2025-10-08 16:45:50.301929759 +0000 UTC m=+5559.655844660" watchObservedRunningTime="2025-10-08 16:45:50.31389097 +0000 UTC m=+5559.667805871" Oct 08 16:45:52 crc kubenswrapper[4945]: I1008 16:45:52.517908 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:52 crc kubenswrapper[4945]: I1008 16:45:52.519364 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:45:53 crc kubenswrapper[4945]: I1008 16:45:53.564720 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hxdzq" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="registry-server" probeResult="failure" output=< Oct 08 16:45:53 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:45:53 crc kubenswrapper[4945]: > Oct 08 16:46:03 crc kubenswrapper[4945]: I1008 16:46:03.574453 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hxdzq" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="registry-server" probeResult="failure" output=< Oct 08 16:46:03 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:46:03 crc kubenswrapper[4945]: > Oct 08 16:46:13 crc kubenswrapper[4945]: I1008 16:46:13.561925 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hxdzq" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="registry-server" probeResult="failure" output=< Oct 08 16:46:13 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:46:13 crc kubenswrapper[4945]: > Oct 08 16:46:22 crc kubenswrapper[4945]: I1008 16:46:22.576117 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:46:22 crc kubenswrapper[4945]: I1008 16:46:22.641650 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:46:22 crc kubenswrapper[4945]: I1008 16:46:22.820272 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzq"] Oct 08 16:46:24 crc kubenswrapper[4945]: I1008 16:46:24.600619 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hxdzq" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="registry-server" containerID="cri-o://f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3" gracePeriod=2 Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.104796 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.248144 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-utilities\") pod \"03117354-3316-41c5-b997-b24a64f6643e\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.248188 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-catalog-content\") pod \"03117354-3316-41c5-b997-b24a64f6643e\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.248243 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pbxd\" (UniqueName: \"kubernetes.io/projected/03117354-3316-41c5-b997-b24a64f6643e-kube-api-access-5pbxd\") pod \"03117354-3316-41c5-b997-b24a64f6643e\" (UID: \"03117354-3316-41c5-b997-b24a64f6643e\") " Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.249072 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-utilities" (OuterVolumeSpecName: "utilities") pod "03117354-3316-41c5-b997-b24a64f6643e" (UID: "03117354-3316-41c5-b997-b24a64f6643e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.264315 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03117354-3316-41c5-b997-b24a64f6643e-kube-api-access-5pbxd" (OuterVolumeSpecName: "kube-api-access-5pbxd") pod "03117354-3316-41c5-b997-b24a64f6643e" (UID: "03117354-3316-41c5-b997-b24a64f6643e"). InnerVolumeSpecName "kube-api-access-5pbxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.317935 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03117354-3316-41c5-b997-b24a64f6643e" (UID: "03117354-3316-41c5-b997-b24a64f6643e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.350893 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.351153 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03117354-3316-41c5-b997-b24a64f6643e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.351229 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pbxd\" (UniqueName: \"kubernetes.io/projected/03117354-3316-41c5-b997-b24a64f6643e-kube-api-access-5pbxd\") on node \"crc\" DevicePath \"\"" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.611069 4945 generic.go:334] "Generic (PLEG): container finished" podID="03117354-3316-41c5-b997-b24a64f6643e" containerID="f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3" exitCode=0 Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.611133 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzq" event={"ID":"03117354-3316-41c5-b997-b24a64f6643e","Type":"ContainerDied","Data":"f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3"} Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.611432 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzq" event={"ID":"03117354-3316-41c5-b997-b24a64f6643e","Type":"ContainerDied","Data":"8ae683a4bccc3613100d008cc7bf4d80052324ca8c7540d27e51aba85afa9d05"} Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.611454 4945 scope.go:117] "RemoveContainer" containerID="f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.611176 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzq" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.638399 4945 scope.go:117] "RemoveContainer" containerID="78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.650023 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzq"] Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.658823 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzq"] Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.664472 4945 scope.go:117] "RemoveContainer" containerID="5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.724888 4945 scope.go:117] "RemoveContainer" containerID="f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3" Oct 08 16:46:25 crc kubenswrapper[4945]: E1008 16:46:25.725370 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3\": container with ID starting with f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3 not found: ID does not exist" containerID="f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.725412 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3"} err="failed to get container status \"f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3\": rpc error: code = NotFound desc = could not find container \"f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3\": container with ID starting with f78aacabb745ded3417c791580e24762a7af2acee599a1a3aa0192c75ece07b3 not found: ID does not exist" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.725433 4945 scope.go:117] "RemoveContainer" containerID="78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463" Oct 08 16:46:25 crc kubenswrapper[4945]: E1008 16:46:25.725801 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463\": container with ID starting with 78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463 not found: ID does not exist" containerID="78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.725827 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463"} err="failed to get container status \"78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463\": rpc error: code = NotFound desc = could not find container \"78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463\": container with ID starting with 78ddb502c723b381365336d82668b9903e17cd2097844a106e2c86c31523f463 not found: ID does not exist" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.725849 4945 scope.go:117] "RemoveContainer" containerID="5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007" Oct 08 16:46:25 crc kubenswrapper[4945]: E1008 16:46:25.726235 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007\": container with ID starting with 5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007 not found: ID does not exist" containerID="5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007" Oct 08 16:46:25 crc kubenswrapper[4945]: I1008 16:46:25.726259 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007"} err="failed to get container status \"5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007\": rpc error: code = NotFound desc = could not find container \"5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007\": container with ID starting with 5d2bfb97edca8d4285de1fd8b2e959a28c463364f28df50aa2ad8b983db3a007 not found: ID does not exist" Oct 08 16:46:26 crc kubenswrapper[4945]: I1008 16:46:26.037685 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03117354-3316-41c5-b997-b24a64f6643e" path="/var/lib/kubelet/pods/03117354-3316-41c5-b997-b24a64f6643e/volumes" Oct 08 16:47:01 crc kubenswrapper[4945]: I1008 16:47:01.916696 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gdfmc"] Oct 08 16:47:01 crc kubenswrapper[4945]: E1008 16:47:01.917719 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="registry-server" Oct 08 16:47:01 crc kubenswrapper[4945]: I1008 16:47:01.917738 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="registry-server" Oct 08 16:47:01 crc kubenswrapper[4945]: E1008 16:47:01.917772 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="extract-utilities" Oct 08 16:47:01 crc kubenswrapper[4945]: I1008 16:47:01.917781 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="extract-utilities" Oct 08 16:47:01 crc kubenswrapper[4945]: E1008 16:47:01.917798 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="extract-content" Oct 08 16:47:01 crc kubenswrapper[4945]: I1008 16:47:01.917807 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="extract-content" Oct 08 16:47:01 crc kubenswrapper[4945]: I1008 16:47:01.918066 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="03117354-3316-41c5-b997-b24a64f6643e" containerName="registry-server" Oct 08 16:47:01 crc kubenswrapper[4945]: I1008 16:47:01.920454 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:01 crc kubenswrapper[4945]: I1008 16:47:01.954621 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gdfmc"] Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.030014 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvxbt\" (UniqueName: \"kubernetes.io/projected/8bd6ada4-31a8-4983-8545-026cb5e003cd-kube-api-access-kvxbt\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.030055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-catalog-content\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.030252 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-utilities\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.131837 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-utilities\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.131954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvxbt\" (UniqueName: \"kubernetes.io/projected/8bd6ada4-31a8-4983-8545-026cb5e003cd-kube-api-access-kvxbt\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.131983 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-catalog-content\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.132425 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-utilities\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.132469 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-catalog-content\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.151142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvxbt\" (UniqueName: \"kubernetes.io/projected/8bd6ada4-31a8-4983-8545-026cb5e003cd-kube-api-access-kvxbt\") pod \"redhat-marketplace-gdfmc\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.256727 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.773209 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gdfmc"] Oct 08 16:47:02 crc kubenswrapper[4945]: I1008 16:47:02.974245 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdfmc" event={"ID":"8bd6ada4-31a8-4983-8545-026cb5e003cd","Type":"ContainerStarted","Data":"cd2fe9204b3eca5f31d995f9dd44f7932396572bef8061f72866add2a14b8cbc"} Oct 08 16:47:03 crc kubenswrapper[4945]: I1008 16:47:03.988063 4945 generic.go:334] "Generic (PLEG): container finished" podID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerID="2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e" exitCode=0 Oct 08 16:47:03 crc kubenswrapper[4945]: I1008 16:47:03.988154 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdfmc" event={"ID":"8bd6ada4-31a8-4983-8545-026cb5e003cd","Type":"ContainerDied","Data":"2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e"} Oct 08 16:47:05 crc kubenswrapper[4945]: I1008 16:47:05.005627 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdfmc" event={"ID":"8bd6ada4-31a8-4983-8545-026cb5e003cd","Type":"ContainerStarted","Data":"dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69"} Oct 08 16:47:06 crc kubenswrapper[4945]: I1008 16:47:06.014694 4945 generic.go:334] "Generic (PLEG): container finished" podID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerID="dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69" exitCode=0 Oct 08 16:47:06 crc kubenswrapper[4945]: I1008 16:47:06.014738 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdfmc" event={"ID":"8bd6ada4-31a8-4983-8545-026cb5e003cd","Type":"ContainerDied","Data":"dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69"} Oct 08 16:47:07 crc kubenswrapper[4945]: I1008 16:47:07.034771 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdfmc" event={"ID":"8bd6ada4-31a8-4983-8545-026cb5e003cd","Type":"ContainerStarted","Data":"2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529"} Oct 08 16:47:07 crc kubenswrapper[4945]: I1008 16:47:07.065775 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gdfmc" podStartSLOduration=3.637643076 podStartE2EDuration="6.06575734s" podCreationTimestamp="2025-10-08 16:47:01 +0000 UTC" firstStartedPulling="2025-10-08 16:47:03.990684958 +0000 UTC m=+5633.344599869" lastFinishedPulling="2025-10-08 16:47:06.418799232 +0000 UTC m=+5635.772714133" observedRunningTime="2025-10-08 16:47:07.058462856 +0000 UTC m=+5636.412377757" watchObservedRunningTime="2025-10-08 16:47:07.06575734 +0000 UTC m=+5636.419672241" Oct 08 16:47:12 crc kubenswrapper[4945]: I1008 16:47:12.257498 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:12 crc kubenswrapper[4945]: I1008 16:47:12.258021 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:12 crc kubenswrapper[4945]: I1008 16:47:12.310867 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:13 crc kubenswrapper[4945]: I1008 16:47:13.135688 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:13 crc kubenswrapper[4945]: I1008 16:47:13.185577 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gdfmc"] Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.107282 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gdfmc" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="registry-server" containerID="cri-o://2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529" gracePeriod=2 Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.652145 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.840908 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-utilities\") pod \"8bd6ada4-31a8-4983-8545-026cb5e003cd\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.840955 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-catalog-content\") pod \"8bd6ada4-31a8-4983-8545-026cb5e003cd\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.840985 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvxbt\" (UniqueName: \"kubernetes.io/projected/8bd6ada4-31a8-4983-8545-026cb5e003cd-kube-api-access-kvxbt\") pod \"8bd6ada4-31a8-4983-8545-026cb5e003cd\" (UID: \"8bd6ada4-31a8-4983-8545-026cb5e003cd\") " Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.841637 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-utilities" (OuterVolumeSpecName: "utilities") pod "8bd6ada4-31a8-4983-8545-026cb5e003cd" (UID: "8bd6ada4-31a8-4983-8545-026cb5e003cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.853869 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bd6ada4-31a8-4983-8545-026cb5e003cd" (UID: "8bd6ada4-31a8-4983-8545-026cb5e003cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.858056 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bd6ada4-31a8-4983-8545-026cb5e003cd-kube-api-access-kvxbt" (OuterVolumeSpecName: "kube-api-access-kvxbt") pod "8bd6ada4-31a8-4983-8545-026cb5e003cd" (UID: "8bd6ada4-31a8-4983-8545-026cb5e003cd"). InnerVolumeSpecName "kube-api-access-kvxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.943075 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.943128 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd6ada4-31a8-4983-8545-026cb5e003cd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:47:15 crc kubenswrapper[4945]: I1008 16:47:15.943144 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvxbt\" (UniqueName: \"kubernetes.io/projected/8bd6ada4-31a8-4983-8545-026cb5e003cd-kube-api-access-kvxbt\") on node \"crc\" DevicePath \"\"" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.119544 4945 generic.go:334] "Generic (PLEG): container finished" podID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerID="2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529" exitCode=0 Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.119593 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gdfmc" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.119626 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdfmc" event={"ID":"8bd6ada4-31a8-4983-8545-026cb5e003cd","Type":"ContainerDied","Data":"2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529"} Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.119977 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdfmc" event={"ID":"8bd6ada4-31a8-4983-8545-026cb5e003cd","Type":"ContainerDied","Data":"cd2fe9204b3eca5f31d995f9dd44f7932396572bef8061f72866add2a14b8cbc"} Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.120000 4945 scope.go:117] "RemoveContainer" containerID="2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.151898 4945 scope.go:117] "RemoveContainer" containerID="dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.152642 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gdfmc"] Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.164645 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gdfmc"] Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.183209 4945 scope.go:117] "RemoveContainer" containerID="2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.226638 4945 scope.go:117] "RemoveContainer" containerID="2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529" Oct 08 16:47:16 crc kubenswrapper[4945]: E1008 16:47:16.227169 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529\": container with ID starting with 2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529 not found: ID does not exist" containerID="2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.227222 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529"} err="failed to get container status \"2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529\": rpc error: code = NotFound desc = could not find container \"2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529\": container with ID starting with 2f3d21387f31137843d8a540ce2461bbbc661d8375489388781cde845bc64529 not found: ID does not exist" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.227254 4945 scope.go:117] "RemoveContainer" containerID="dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69" Oct 08 16:47:16 crc kubenswrapper[4945]: E1008 16:47:16.227528 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69\": container with ID starting with dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69 not found: ID does not exist" containerID="dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.227554 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69"} err="failed to get container status \"dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69\": rpc error: code = NotFound desc = could not find container \"dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69\": container with ID starting with dac6785adf15721482f66bc0c31fe4d6126b6a5687b1720213bec63943637e69 not found: ID does not exist" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.227569 4945 scope.go:117] "RemoveContainer" containerID="2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e" Oct 08 16:47:16 crc kubenswrapper[4945]: E1008 16:47:16.227785 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e\": container with ID starting with 2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e not found: ID does not exist" containerID="2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e" Oct 08 16:47:16 crc kubenswrapper[4945]: I1008 16:47:16.227807 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e"} err="failed to get container status \"2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e\": rpc error: code = NotFound desc = could not find container \"2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e\": container with ID starting with 2ee0e57f81a5225353cde75ac9d5317fa1d1c4627118f9653a5c290d2d63164e not found: ID does not exist" Oct 08 16:47:18 crc kubenswrapper[4945]: I1008 16:47:18.035480 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" path="/var/lib/kubelet/pods/8bd6ada4-31a8-4983-8545-026cb5e003cd/volumes" Oct 08 16:47:49 crc kubenswrapper[4945]: I1008 16:47:49.184968 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:47:49 crc kubenswrapper[4945]: I1008 16:47:49.185549 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.260339 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c4pcq"] Oct 08 16:47:55 crc kubenswrapper[4945]: E1008 16:47:55.261416 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="extract-utilities" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.261434 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="extract-utilities" Oct 08 16:47:55 crc kubenswrapper[4945]: E1008 16:47:55.261445 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="registry-server" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.261453 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="registry-server" Oct 08 16:47:55 crc kubenswrapper[4945]: E1008 16:47:55.261498 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="extract-content" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.261507 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="extract-content" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.261770 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bd6ada4-31a8-4983-8545-026cb5e003cd" containerName="registry-server" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.263537 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.288908 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c4pcq"] Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.397649 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-catalog-content\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.398010 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8njjc\" (UniqueName: \"kubernetes.io/projected/5017c5a6-6183-471e-9da3-f5c1f02c9898-kube-api-access-8njjc\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.398208 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-utilities\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.500058 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-catalog-content\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.500196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8njjc\" (UniqueName: \"kubernetes.io/projected/5017c5a6-6183-471e-9da3-f5c1f02c9898-kube-api-access-8njjc\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.500295 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-utilities\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.500861 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-utilities\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.500928 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-catalog-content\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.521806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8njjc\" (UniqueName: \"kubernetes.io/projected/5017c5a6-6183-471e-9da3-f5c1f02c9898-kube-api-access-8njjc\") pod \"community-operators-c4pcq\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:55 crc kubenswrapper[4945]: I1008 16:47:55.590864 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:47:56 crc kubenswrapper[4945]: I1008 16:47:56.088697 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c4pcq"] Oct 08 16:47:56 crc kubenswrapper[4945]: I1008 16:47:56.497920 4945 generic.go:334] "Generic (PLEG): container finished" podID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerID="9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4" exitCode=0 Oct 08 16:47:56 crc kubenswrapper[4945]: I1008 16:47:56.498032 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pcq" event={"ID":"5017c5a6-6183-471e-9da3-f5c1f02c9898","Type":"ContainerDied","Data":"9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4"} Oct 08 16:47:56 crc kubenswrapper[4945]: I1008 16:47:56.498339 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pcq" event={"ID":"5017c5a6-6183-471e-9da3-f5c1f02c9898","Type":"ContainerStarted","Data":"61a69ef819eaf722d0e7999affdfe05bc70306f656c497650c6d309d3fb667a3"} Oct 08 16:47:58 crc kubenswrapper[4945]: I1008 16:47:58.519092 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pcq" event={"ID":"5017c5a6-6183-471e-9da3-f5c1f02c9898","Type":"ContainerStarted","Data":"31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770"} Oct 08 16:47:59 crc kubenswrapper[4945]: I1008 16:47:59.534170 4945 generic.go:334] "Generic (PLEG): container finished" podID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerID="31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770" exitCode=0 Oct 08 16:47:59 crc kubenswrapper[4945]: I1008 16:47:59.534228 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pcq" event={"ID":"5017c5a6-6183-471e-9da3-f5c1f02c9898","Type":"ContainerDied","Data":"31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770"} Oct 08 16:48:00 crc kubenswrapper[4945]: I1008 16:48:00.545278 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pcq" event={"ID":"5017c5a6-6183-471e-9da3-f5c1f02c9898","Type":"ContainerStarted","Data":"979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20"} Oct 08 16:48:00 crc kubenswrapper[4945]: I1008 16:48:00.564998 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c4pcq" podStartSLOduration=2.141118927 podStartE2EDuration="5.564981666s" podCreationTimestamp="2025-10-08 16:47:55 +0000 UTC" firstStartedPulling="2025-10-08 16:47:56.500089333 +0000 UTC m=+5685.854004234" lastFinishedPulling="2025-10-08 16:47:59.923952072 +0000 UTC m=+5689.277866973" observedRunningTime="2025-10-08 16:48:00.562893834 +0000 UTC m=+5689.916808735" watchObservedRunningTime="2025-10-08 16:48:00.564981666 +0000 UTC m=+5689.918896567" Oct 08 16:48:05 crc kubenswrapper[4945]: I1008 16:48:05.591395 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:48:05 crc kubenswrapper[4945]: I1008 16:48:05.592241 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:48:06 crc kubenswrapper[4945]: I1008 16:48:06.648343 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-c4pcq" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="registry-server" probeResult="failure" output=< Oct 08 16:48:06 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 08 16:48:06 crc kubenswrapper[4945]: > Oct 08 16:48:15 crc kubenswrapper[4945]: I1008 16:48:15.636848 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:48:15 crc kubenswrapper[4945]: I1008 16:48:15.685652 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:48:15 crc kubenswrapper[4945]: I1008 16:48:15.879381 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c4pcq"] Oct 08 16:48:16 crc kubenswrapper[4945]: I1008 16:48:16.716659 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c4pcq" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="registry-server" containerID="cri-o://979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20" gracePeriod=2 Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.196632 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.381793 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8njjc\" (UniqueName: \"kubernetes.io/projected/5017c5a6-6183-471e-9da3-f5c1f02c9898-kube-api-access-8njjc\") pod \"5017c5a6-6183-471e-9da3-f5c1f02c9898\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.381855 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-catalog-content\") pod \"5017c5a6-6183-471e-9da3-f5c1f02c9898\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.381938 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-utilities\") pod \"5017c5a6-6183-471e-9da3-f5c1f02c9898\" (UID: \"5017c5a6-6183-471e-9da3-f5c1f02c9898\") " Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.383014 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-utilities" (OuterVolumeSpecName: "utilities") pod "5017c5a6-6183-471e-9da3-f5c1f02c9898" (UID: "5017c5a6-6183-471e-9da3-f5c1f02c9898"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.413155 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5017c5a6-6183-471e-9da3-f5c1f02c9898-kube-api-access-8njjc" (OuterVolumeSpecName: "kube-api-access-8njjc") pod "5017c5a6-6183-471e-9da3-f5c1f02c9898" (UID: "5017c5a6-6183-471e-9da3-f5c1f02c9898"). InnerVolumeSpecName "kube-api-access-8njjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.430627 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5017c5a6-6183-471e-9da3-f5c1f02c9898" (UID: "5017c5a6-6183-471e-9da3-f5c1f02c9898"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.484598 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.484634 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5017c5a6-6183-471e-9da3-f5c1f02c9898-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.484664 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8njjc\" (UniqueName: \"kubernetes.io/projected/5017c5a6-6183-471e-9da3-f5c1f02c9898-kube-api-access-8njjc\") on node \"crc\" DevicePath \"\"" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.730080 4945 generic.go:334] "Generic (PLEG): container finished" podID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerID="979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20" exitCode=0 Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.730147 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pcq" event={"ID":"5017c5a6-6183-471e-9da3-f5c1f02c9898","Type":"ContainerDied","Data":"979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20"} Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.730178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pcq" event={"ID":"5017c5a6-6183-471e-9da3-f5c1f02c9898","Type":"ContainerDied","Data":"61a69ef819eaf722d0e7999affdfe05bc70306f656c497650c6d309d3fb667a3"} Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.730182 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pcq" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.730198 4945 scope.go:117] "RemoveContainer" containerID="979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.752042 4945 scope.go:117] "RemoveContainer" containerID="31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.780494 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c4pcq"] Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.791373 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c4pcq"] Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.803841 4945 scope.go:117] "RemoveContainer" containerID="9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.833945 4945 scope.go:117] "RemoveContainer" containerID="979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20" Oct 08 16:48:17 crc kubenswrapper[4945]: E1008 16:48:17.834328 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20\": container with ID starting with 979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20 not found: ID does not exist" containerID="979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.834355 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20"} err="failed to get container status \"979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20\": rpc error: code = NotFound desc = could not find container \"979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20\": container with ID starting with 979d75b2c9f2261065b56b07d0676f5d243f4a4f14ac966f646745d0429bad20 not found: ID does not exist" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.834375 4945 scope.go:117] "RemoveContainer" containerID="31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770" Oct 08 16:48:17 crc kubenswrapper[4945]: E1008 16:48:17.834798 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770\": container with ID starting with 31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770 not found: ID does not exist" containerID="31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.834855 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770"} err="failed to get container status \"31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770\": rpc error: code = NotFound desc = could not find container \"31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770\": container with ID starting with 31185c97f619336b8a318c6aa25eab3bd97d418dc18128d9fedea51d02332770 not found: ID does not exist" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.834882 4945 scope.go:117] "RemoveContainer" containerID="9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4" Oct 08 16:48:17 crc kubenswrapper[4945]: E1008 16:48:17.835399 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4\": container with ID starting with 9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4 not found: ID does not exist" containerID="9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4" Oct 08 16:48:17 crc kubenswrapper[4945]: I1008 16:48:17.835425 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4"} err="failed to get container status \"9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4\": rpc error: code = NotFound desc = could not find container \"9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4\": container with ID starting with 9a6dcbb9a4423609b6374952807068ef87a444924157a39b88f0e9ddb70372a4 not found: ID does not exist" Oct 08 16:48:18 crc kubenswrapper[4945]: I1008 16:48:18.039920 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" path="/var/lib/kubelet/pods/5017c5a6-6183-471e-9da3-f5c1f02c9898/volumes" Oct 08 16:48:19 crc kubenswrapper[4945]: I1008 16:48:19.184189 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:48:19 crc kubenswrapper[4945]: I1008 16:48:19.184545 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:48:49 crc kubenswrapper[4945]: I1008 16:48:49.184786 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:48:49 crc kubenswrapper[4945]: I1008 16:48:49.185465 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:48:49 crc kubenswrapper[4945]: I1008 16:48:49.185510 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:48:49 crc kubenswrapper[4945]: I1008 16:48:49.186421 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:48:49 crc kubenswrapper[4945]: I1008 16:48:49.186483 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" gracePeriod=600 Oct 08 16:48:49 crc kubenswrapper[4945]: E1008 16:48:49.310518 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:48:50 crc kubenswrapper[4945]: I1008 16:48:50.081833 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" exitCode=0 Oct 08 16:48:50 crc kubenswrapper[4945]: I1008 16:48:50.081904 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1"} Oct 08 16:48:50 crc kubenswrapper[4945]: I1008 16:48:50.082190 4945 scope.go:117] "RemoveContainer" containerID="a9568695c8a4dca15d9f95043c6fb8f3e6a9782b4a613574a9f0241ad503097e" Oct 08 16:48:50 crc kubenswrapper[4945]: I1008 16:48:50.082945 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:48:50 crc kubenswrapper[4945]: E1008 16:48:50.084186 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:49:02 crc kubenswrapper[4945]: I1008 16:49:02.206259 4945 generic.go:334] "Generic (PLEG): container finished" podID="0f3aa285-da71-4df6-8188-cc528008a61b" containerID="6b38deae846260c42fa56914bb7126d29bb4c4a9a83e1a25e314279899d5cfb1" exitCode=0 Oct 08 16:49:02 crc kubenswrapper[4945]: I1008 16:49:02.206847 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0f3aa285-da71-4df6-8188-cc528008a61b","Type":"ContainerDied","Data":"6b38deae846260c42fa56914bb7126d29bb4c4a9a83e1a25e314279899d5cfb1"} Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.024402 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:49:03 crc kubenswrapper[4945]: E1008 16:49:03.024730 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.561551 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662488 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config-secret\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662599 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-workdir\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662621 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662676 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-config-data\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662707 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ca-certs\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662722 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ssh-key\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662760 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xjck\" (UniqueName: \"kubernetes.io/projected/0f3aa285-da71-4df6-8188-cc528008a61b-kube-api-access-8xjck\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662805 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-temporary\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.662857 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0f3aa285-da71-4df6-8188-cc528008a61b\" (UID: \"0f3aa285-da71-4df6-8188-cc528008a61b\") " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.663588 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.663728 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-config-data" (OuterVolumeSpecName: "config-data") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.666560 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.681958 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.682629 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f3aa285-da71-4df6-8188-cc528008a61b-kube-api-access-8xjck" (OuterVolumeSpecName: "kube-api-access-8xjck") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "kube-api-access-8xjck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.698839 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.702060 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.703088 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.727181 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0f3aa285-da71-4df6-8188-cc528008a61b" (UID: "0f3aa285-da71-4df6-8188-cc528008a61b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765701 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xjck\" (UniqueName: \"kubernetes.io/projected/0f3aa285-da71-4df6-8188-cc528008a61b-kube-api-access-8xjck\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765752 4945 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765795 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765810 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765825 4945 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765836 4945 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0f3aa285-da71-4df6-8188-cc528008a61b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765848 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f3aa285-da71-4df6-8188-cc528008a61b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765858 4945 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.765867 4945 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f3aa285-da71-4df6-8188-cc528008a61b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.787428 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 16:49:03 crc kubenswrapper[4945]: I1008 16:49:03.868283 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 16:49:04 crc kubenswrapper[4945]: I1008 16:49:04.227453 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0f3aa285-da71-4df6-8188-cc528008a61b","Type":"ContainerDied","Data":"59f70af86fe4980a8e7c8c09d07ff378c2e1a936fbd3e6ce91632ac00668e5f2"} Oct 08 16:49:04 crc kubenswrapper[4945]: I1008 16:49:04.227493 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59f70af86fe4980a8e7c8c09d07ff378c2e1a936fbd3e6ce91632ac00668e5f2" Oct 08 16:49:04 crc kubenswrapper[4945]: I1008 16:49:04.227511 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.786713 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 16:49:06 crc kubenswrapper[4945]: E1008 16:49:06.787626 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f3aa285-da71-4df6-8188-cc528008a61b" containerName="tempest-tests-tempest-tests-runner" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.787643 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f3aa285-da71-4df6-8188-cc528008a61b" containerName="tempest-tests-tempest-tests-runner" Oct 08 16:49:06 crc kubenswrapper[4945]: E1008 16:49:06.787661 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="extract-utilities" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.787669 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="extract-utilities" Oct 08 16:49:06 crc kubenswrapper[4945]: E1008 16:49:06.787683 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="extract-content" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.787690 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="extract-content" Oct 08 16:49:06 crc kubenswrapper[4945]: E1008 16:49:06.787713 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="registry-server" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.787719 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="registry-server" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.787940 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5017c5a6-6183-471e-9da3-f5c1f02c9898" containerName="registry-server" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.787963 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f3aa285-da71-4df6-8188-cc528008a61b" containerName="tempest-tests-tempest-tests-runner" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.788747 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.790929 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mwxqb" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.802668 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.936062 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"af253961-462b-4f6a-95dc-088795589466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:06 crc kubenswrapper[4945]: I1008 16:49:06.936170 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx8zj\" (UniqueName: \"kubernetes.io/projected/af253961-462b-4f6a-95dc-088795589466-kube-api-access-fx8zj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"af253961-462b-4f6a-95dc-088795589466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:07 crc kubenswrapper[4945]: I1008 16:49:07.037668 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"af253961-462b-4f6a-95dc-088795589466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:07 crc kubenswrapper[4945]: I1008 16:49:07.037762 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx8zj\" (UniqueName: \"kubernetes.io/projected/af253961-462b-4f6a-95dc-088795589466-kube-api-access-fx8zj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"af253961-462b-4f6a-95dc-088795589466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:07 crc kubenswrapper[4945]: I1008 16:49:07.038416 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"af253961-462b-4f6a-95dc-088795589466\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:07 crc kubenswrapper[4945]: I1008 16:49:07.075132 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx8zj\" (UniqueName: \"kubernetes.io/projected/af253961-462b-4f6a-95dc-088795589466-kube-api-access-fx8zj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"af253961-462b-4f6a-95dc-088795589466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:07 crc kubenswrapper[4945]: I1008 16:49:07.091232 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"af253961-462b-4f6a-95dc-088795589466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:07 crc kubenswrapper[4945]: I1008 16:49:07.114605 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 16:49:07 crc kubenswrapper[4945]: I1008 16:49:07.578661 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 16:49:08 crc kubenswrapper[4945]: I1008 16:49:08.271534 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"af253961-462b-4f6a-95dc-088795589466","Type":"ContainerStarted","Data":"4871a0c7862cbf9c67048fdb17c40655d8ab05efc2fa83d18ca8d242d80ecf45"} Oct 08 16:49:09 crc kubenswrapper[4945]: I1008 16:49:09.283526 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"af253961-462b-4f6a-95dc-088795589466","Type":"ContainerStarted","Data":"468e055059a0eb560be2686ea7de55422cbff76b89840cd9279eccac247e7deb"} Oct 08 16:49:09 crc kubenswrapper[4945]: I1008 16:49:09.305522 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.440849194 podStartE2EDuration="3.305502733s" podCreationTimestamp="2025-10-08 16:49:06 +0000 UTC" firstStartedPulling="2025-10-08 16:49:07.577425769 +0000 UTC m=+5756.931340670" lastFinishedPulling="2025-10-08 16:49:08.442079308 +0000 UTC m=+5757.795994209" observedRunningTime="2025-10-08 16:49:09.297480873 +0000 UTC m=+5758.651395814" watchObservedRunningTime="2025-10-08 16:49:09.305502733 +0000 UTC m=+5758.659417634" Oct 08 16:49:16 crc kubenswrapper[4945]: I1008 16:49:16.025237 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:49:16 crc kubenswrapper[4945]: E1008 16:49:16.026328 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.386063 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lzfw/must-gather-j8dz7"] Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.388667 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.391931 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7lzfw"/"default-dockercfg-8x42h" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.392309 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7lzfw"/"kube-root-ca.crt" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.392527 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7lzfw"/"openshift-service-ca.crt" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.404510 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7lzfw/must-gather-j8dz7"] Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.487094 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l4nm\" (UniqueName: \"kubernetes.io/projected/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-kube-api-access-9l4nm\") pod \"must-gather-j8dz7\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.487370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-must-gather-output\") pod \"must-gather-j8dz7\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.589378 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l4nm\" (UniqueName: \"kubernetes.io/projected/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-kube-api-access-9l4nm\") pod \"must-gather-j8dz7\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.589489 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-must-gather-output\") pod \"must-gather-j8dz7\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.590010 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-must-gather-output\") pod \"must-gather-j8dz7\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.607657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l4nm\" (UniqueName: \"kubernetes.io/projected/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-kube-api-access-9l4nm\") pod \"must-gather-j8dz7\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:28 crc kubenswrapper[4945]: I1008 16:49:28.709731 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:49:29 crc kubenswrapper[4945]: W1008 16:49:29.178350 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad7b6960_2f9c_49cc_bbc4_af23dfcaf0a5.slice/crio-a90693d7c6f3d0c232579e9384c7cb05fd56a0c717bef1f08f21d953ee12d8cd WatchSource:0}: Error finding container a90693d7c6f3d0c232579e9384c7cb05fd56a0c717bef1f08f21d953ee12d8cd: Status 404 returned error can't find the container with id a90693d7c6f3d0c232579e9384c7cb05fd56a0c717bef1f08f21d953ee12d8cd Oct 08 16:49:29 crc kubenswrapper[4945]: I1008 16:49:29.180099 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7lzfw/must-gather-j8dz7"] Oct 08 16:49:29 crc kubenswrapper[4945]: I1008 16:49:29.482133 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" event={"ID":"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5","Type":"ContainerStarted","Data":"a90693d7c6f3d0c232579e9384c7cb05fd56a0c717bef1f08f21d953ee12d8cd"} Oct 08 16:49:31 crc kubenswrapper[4945]: I1008 16:49:31.024169 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:49:31 crc kubenswrapper[4945]: E1008 16:49:31.024790 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:49:37 crc kubenswrapper[4945]: I1008 16:49:37.561090 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" event={"ID":"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5","Type":"ContainerStarted","Data":"120f101da0c28389cb8d9753e1475870d26f0e0233bff15618ea04dd01c79ace"} Oct 08 16:49:37 crc kubenswrapper[4945]: I1008 16:49:37.561770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" event={"ID":"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5","Type":"ContainerStarted","Data":"989abd9d6a3b49c6c2c43ddc5c4a6e032d727076a4aa8187fc703d10c4878b91"} Oct 08 16:49:37 crc kubenswrapper[4945]: I1008 16:49:37.586510 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" podStartSLOduration=1.927007786 podStartE2EDuration="9.58648934s" podCreationTimestamp="2025-10-08 16:49:28 +0000 UTC" firstStartedPulling="2025-10-08 16:49:29.180670673 +0000 UTC m=+5778.534585574" lastFinishedPulling="2025-10-08 16:49:36.840152227 +0000 UTC m=+5786.194067128" observedRunningTime="2025-10-08 16:49:37.577720511 +0000 UTC m=+5786.931635412" watchObservedRunningTime="2025-10-08 16:49:37.58648934 +0000 UTC m=+5786.940404241" Oct 08 16:49:39 crc kubenswrapper[4945]: E1008 16:49:39.788193 4945 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.138:43466->38.102.83.138:44411: write tcp 38.102.83.138:43466->38.102.83.138:44411: write: broken pipe Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.038573 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-jnm5r"] Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.041319 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.093602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8742445-70b3-474f-b4d0-ce61044ba8c7-host\") pod \"crc-debug-jnm5r\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.094003 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrg62\" (UniqueName: \"kubernetes.io/projected/f8742445-70b3-474f-b4d0-ce61044ba8c7-kube-api-access-lrg62\") pod \"crc-debug-jnm5r\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.196031 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8742445-70b3-474f-b4d0-ce61044ba8c7-host\") pod \"crc-debug-jnm5r\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.196156 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrg62\" (UniqueName: \"kubernetes.io/projected/f8742445-70b3-474f-b4d0-ce61044ba8c7-kube-api-access-lrg62\") pod \"crc-debug-jnm5r\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.196185 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8742445-70b3-474f-b4d0-ce61044ba8c7-host\") pod \"crc-debug-jnm5r\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.219365 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrg62\" (UniqueName: \"kubernetes.io/projected/f8742445-70b3-474f-b4d0-ce61044ba8c7-kube-api-access-lrg62\") pod \"crc-debug-jnm5r\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.364044 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:49:41 crc kubenswrapper[4945]: I1008 16:49:41.600551 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" event={"ID":"f8742445-70b3-474f-b4d0-ce61044ba8c7","Type":"ContainerStarted","Data":"5a810958e4b7f9061858d29b514e759f904edb1eded12d4191083f43a34a42ac"} Oct 08 16:49:42 crc kubenswrapper[4945]: I1008 16:49:42.031232 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:49:42 crc kubenswrapper[4945]: E1008 16:49:42.031871 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:49:52 crc kubenswrapper[4945]: I1008 16:49:52.713938 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" event={"ID":"f8742445-70b3-474f-b4d0-ce61044ba8c7","Type":"ContainerStarted","Data":"4c9b3a7be6c353ced2412ef28494ce7b75f4b2b57673fd9351fd891b6a539bbb"} Oct 08 16:49:52 crc kubenswrapper[4945]: I1008 16:49:52.734872 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" podStartSLOduration=1.6018034719999998 podStartE2EDuration="11.734855172s" podCreationTimestamp="2025-10-08 16:49:41 +0000 UTC" firstStartedPulling="2025-10-08 16:49:41.41530698 +0000 UTC m=+5790.769221881" lastFinishedPulling="2025-10-08 16:49:51.54835865 +0000 UTC m=+5800.902273581" observedRunningTime="2025-10-08 16:49:52.725905389 +0000 UTC m=+5802.079820290" watchObservedRunningTime="2025-10-08 16:49:52.734855172 +0000 UTC m=+5802.088770073" Oct 08 16:49:57 crc kubenswrapper[4945]: I1008 16:49:57.024608 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:49:57 crc kubenswrapper[4945]: E1008 16:49:57.025285 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:50:08 crc kubenswrapper[4945]: I1008 16:50:08.024159 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:50:08 crc kubenswrapper[4945]: E1008 16:50:08.025013 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:50:20 crc kubenswrapper[4945]: I1008 16:50:20.023901 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:50:20 crc kubenswrapper[4945]: E1008 16:50:20.024661 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:50:32 crc kubenswrapper[4945]: I1008 16:50:32.032088 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:50:32 crc kubenswrapper[4945]: E1008 16:50:32.032965 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:50:47 crc kubenswrapper[4945]: I1008 16:50:47.025006 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:50:47 crc kubenswrapper[4945]: E1008 16:50:47.030934 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:50:58 crc kubenswrapper[4945]: I1008 16:50:58.023980 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:50:58 crc kubenswrapper[4945]: E1008 16:50:58.026150 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:51:01 crc kubenswrapper[4945]: I1008 16:51:01.649064 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5588d6b8d6-7nmw2_ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6/barbican-api/0.log" Oct 08 16:51:01 crc kubenswrapper[4945]: I1008 16:51:01.651391 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5588d6b8d6-7nmw2_ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6/barbican-api-log/0.log" Oct 08 16:51:01 crc kubenswrapper[4945]: I1008 16:51:01.898429 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b8d567b6d-zc6xh_66b89671-2c42-4e97-8a74-0e03bac0613a/barbican-keystone-listener/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.001696 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b8d567b6d-zc6xh_66b89671-2c42-4e97-8a74-0e03bac0613a/barbican-keystone-listener-log/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.190842 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fc897d669-lcbrz_dfcd7c98-f494-4669-8e76-995ccb3e768b/barbican-worker/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.250955 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fc897d669-lcbrz_dfcd7c98-f494-4669-8e76-995ccb3e768b/barbican-worker-log/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.433231 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2_de2e46b1-591e-4540-b69e-eee056133019/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.754846 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/ceilometer-notification-agent/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.775982 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/ceilometer-central-agent/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.822510 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/proxy-httpd/0.log" Oct 08 16:51:02 crc kubenswrapper[4945]: I1008 16:51:02.968571 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/sg-core/0.log" Oct 08 16:51:03 crc kubenswrapper[4945]: I1008 16:51:03.219729 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c1bfcac1-b1e7-431f-b145-62a2da972214/cinder-api-log/0.log" Oct 08 16:51:03 crc kubenswrapper[4945]: I1008 16:51:03.404191 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c1bfcac1-b1e7-431f-b145-62a2da972214/cinder-api/0.log" Oct 08 16:51:03 crc kubenswrapper[4945]: I1008 16:51:03.632476 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_a2eda97e-ccd7-42e8-bca5-26402069c250/probe/0.log" Oct 08 16:51:03 crc kubenswrapper[4945]: I1008 16:51:03.633094 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_a2eda97e-ccd7-42e8-bca5-26402069c250/cinder-backup/0.log" Oct 08 16:51:03 crc kubenswrapper[4945]: I1008 16:51:03.851019 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_815a3456-3316-4348-a417-f274fda65c41/cinder-scheduler/0.log" Oct 08 16:51:03 crc kubenswrapper[4945]: I1008 16:51:03.929276 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_815a3456-3316-4348-a417-f274fda65c41/probe/0.log" Oct 08 16:51:04 crc kubenswrapper[4945]: I1008 16:51:04.365563 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_1c0d0627-7d92-4bd7-8f53-3cd9117aec3b/probe/0.log" Oct 08 16:51:04 crc kubenswrapper[4945]: I1008 16:51:04.427672 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_1c0d0627-7d92-4bd7-8f53-3cd9117aec3b/cinder-volume/0.log" Oct 08 16:51:04 crc kubenswrapper[4945]: I1008 16:51:04.724660 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e2f8bf86-ab2b-428a-9e89-2d71091927ef/probe/0.log" Oct 08 16:51:04 crc kubenswrapper[4945]: I1008 16:51:04.730373 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e2f8bf86-ab2b-428a-9e89-2d71091927ef/cinder-volume/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.010519 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t_a6323567-dc30-45ed-8611-4925d983e0ef/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.017432 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh_fcdb8897-6c75-48d4-bbdf-b4eb486331d5/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.285183 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tkrps_71b8f45e-f47d-432d-b22e-3fa39e8059dd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.430518 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f7b68cdfc-bhvjv_108d033e-a880-4b1d-ba80-293d36d0a58a/init/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.615409 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f7b68cdfc-bhvjv_108d033e-a880-4b1d-ba80-293d36d0a58a/init/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.687102 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8_e59943cd-9572-4e6b-8a58-c0ec1283b2aa/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.800496 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f7b68cdfc-bhvjv_108d033e-a880-4b1d-ba80-293d36d0a58a/dnsmasq-dns/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.903924 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bc16ff2a-fe01-42bf-8b16-e4ba10315929/glance-log/0.log" Oct 08 16:51:05 crc kubenswrapper[4945]: I1008 16:51:05.913282 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bc16ff2a-fe01-42bf-8b16-e4ba10315929/glance-httpd/0.log" Oct 08 16:51:06 crc kubenswrapper[4945]: I1008 16:51:06.124749 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f2f0a0ac-d149-4fc1-878a-7e21188cd4ee/glance-log/0.log" Oct 08 16:51:06 crc kubenswrapper[4945]: I1008 16:51:06.147864 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f2f0a0ac-d149-4fc1-878a-7e21188cd4ee/glance-httpd/0.log" Oct 08 16:51:06 crc kubenswrapper[4945]: I1008 16:51:06.465344 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-29p9d_b8610fc4-8948-4247-91b3-78ca4e5e1767/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:06 crc kubenswrapper[4945]: I1008 16:51:06.465857 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c58b7c96-vjh9q_37d6d391-e363-42db-8a0e-4e837e4af4db/horizon/0.log" Oct 08 16:51:06 crc kubenswrapper[4945]: I1008 16:51:06.662443 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-twd5z_bad79416-d5f9-4e87-bab7-cbd0ea8047e9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:07 crc kubenswrapper[4945]: I1008 16:51:07.058067 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332321-qc7kq_ca3e1262-6ad2-4df4-b86d-dc17fc2339c6/keystone-cron/0.log" Oct 08 16:51:07 crc kubenswrapper[4945]: I1008 16:51:07.115784 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c58b7c96-vjh9q_37d6d391-e363-42db-8a0e-4e837e4af4db/horizon-log/0.log" Oct 08 16:51:07 crc kubenswrapper[4945]: I1008 16:51:07.320067 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_803b9d32-2d37-4fd2-96fd-b4001daf26ed/kube-state-metrics/0.log" Oct 08 16:51:07 crc kubenswrapper[4945]: I1008 16:51:07.445831 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-65b767dfc8-l5jkz_f748658a-398b-4b6e-9f27-d2a45e6f6827/keystone-api/0.log" Oct 08 16:51:07 crc kubenswrapper[4945]: I1008 16:51:07.493179 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-76mxw_8a77f15d-769f-493c-b410-eb13c9df8fe6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:08 crc kubenswrapper[4945]: I1008 16:51:08.065917 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5988468485-ctpxh_acfb96c2-ab3d-4e3f-86ce-c6befd371617/neutron-httpd/0.log" Oct 08 16:51:08 crc kubenswrapper[4945]: I1008 16:51:08.107190 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5988468485-ctpxh_acfb96c2-ab3d-4e3f-86ce-c6befd371617/neutron-api/0.log" Oct 08 16:51:08 crc kubenswrapper[4945]: I1008 16:51:08.304452 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p_0c384ac4-a16d-4d25-b5e0-dc7eb33f7352/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:09 crc kubenswrapper[4945]: I1008 16:51:09.289678 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d1cf9a55-81f7-45e6-97b6-b05f4e7f07be/nova-cell0-conductor-conductor/0.log" Oct 08 16:51:09 crc kubenswrapper[4945]: I1008 16:51:09.993894 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ae15d968-e006-4d49-92aa-97cabf627769/nova-cell1-conductor-conductor/0.log" Oct 08 16:51:10 crc kubenswrapper[4945]: I1008 16:51:10.025194 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:51:10 crc kubenswrapper[4945]: E1008 16:51:10.025492 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:51:10 crc kubenswrapper[4945]: I1008 16:51:10.259322 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e8e508c6-7afc-4ae0-84bb-48dece116b1d/nova-api-log/0.log" Oct 08 16:51:10 crc kubenswrapper[4945]: I1008 16:51:10.342223 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e8e508c6-7afc-4ae0-84bb-48dece116b1d/nova-api-api/0.log" Oct 08 16:51:10 crc kubenswrapper[4945]: I1008 16:51:10.549079 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_7a3edb5a-ae0f-411f-b1f5-a9d591a831b6/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 16:51:10 crc kubenswrapper[4945]: I1008 16:51:10.657102 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-mpr75_8b121934-fc40-487d-91fe-cdc91272f2dd/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:10 crc kubenswrapper[4945]: I1008 16:51:10.989985 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1ce5750e-a063-42a1-bde2-0731599fa3bd/nova-metadata-log/0.log" Oct 08 16:51:11 crc kubenswrapper[4945]: I1008 16:51:11.586495 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c17f16f2-5d8f-4178-a6b5-879444c7b262/nova-scheduler-scheduler/0.log" Oct 08 16:51:11 crc kubenswrapper[4945]: I1008 16:51:11.637018 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1d5103f7-4fab-43a6-8989-1b505ac5d8d0/mysql-bootstrap/0.log" Oct 08 16:51:11 crc kubenswrapper[4945]: I1008 16:51:11.813326 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1d5103f7-4fab-43a6-8989-1b505ac5d8d0/mysql-bootstrap/0.log" Oct 08 16:51:11 crc kubenswrapper[4945]: I1008 16:51:11.915480 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1d5103f7-4fab-43a6-8989-1b505ac5d8d0/galera/0.log" Oct 08 16:51:12 crc kubenswrapper[4945]: I1008 16:51:12.181039 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_88f0eb1e-a2bc-4b2f-b14d-8cef0261334e/mysql-bootstrap/0.log" Oct 08 16:51:12 crc kubenswrapper[4945]: I1008 16:51:12.377412 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_88f0eb1e-a2bc-4b2f-b14d-8cef0261334e/mysql-bootstrap/0.log" Oct 08 16:51:12 crc kubenswrapper[4945]: I1008 16:51:12.436173 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_88f0eb1e-a2bc-4b2f-b14d-8cef0261334e/galera/0.log" Oct 08 16:51:12 crc kubenswrapper[4945]: I1008 16:51:12.665382 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_223e741f-c7cc-465f-a4bc-804bfff7a6ef/openstackclient/0.log" Oct 08 16:51:12 crc kubenswrapper[4945]: I1008 16:51:12.884687 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-cwwpc_1b70322d-ae60-46b4-828b-1f4a01cae6b6/ovn-controller/0.log" Oct 08 16:51:13 crc kubenswrapper[4945]: I1008 16:51:13.098735 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-x7dfs_2c0817bf-174f-4d4f-b155-041259384465/openstack-network-exporter/0.log" Oct 08 16:51:13 crc kubenswrapper[4945]: I1008 16:51:13.291785 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1ce5750e-a063-42a1-bde2-0731599fa3bd/nova-metadata-metadata/0.log" Oct 08 16:51:13 crc kubenswrapper[4945]: I1008 16:51:13.334572 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovsdb-server-init/0.log" Oct 08 16:51:13 crc kubenswrapper[4945]: I1008 16:51:13.624985 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovsdb-server-init/0.log" Oct 08 16:51:13 crc kubenswrapper[4945]: I1008 16:51:13.721675 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovsdb-server/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.056459 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovs-vswitchd/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.132862 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-wv9ct_d32f281f-0e3f-4ac5-8463-7bb261bc557e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.321209 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f9854a42-6052-4425-a3a8-190bd5533f75/openstack-network-exporter/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.401604 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f9854a42-6052-4425-a3a8-190bd5533f75/ovn-northd/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.543131 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c4cb5094-ffbc-4fff-bbc4-0a3a35067782/openstack-network-exporter/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.661635 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c4cb5094-ffbc-4fff-bbc4-0a3a35067782/ovsdbserver-nb/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.876488 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec9d8fdb-5e82-4cb6-811b-053a28588453/ovsdbserver-sb/0.log" Oct 08 16:51:14 crc kubenswrapper[4945]: I1008 16:51:14.897897 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec9d8fdb-5e82-4cb6-811b-053a28588453/openstack-network-exporter/0.log" Oct 08 16:51:15 crc kubenswrapper[4945]: I1008 16:51:15.457791 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d5fb75f8-44p9s_bac08d54-03b7-4701-b9c7-69cc37c2eb3d/placement-api/0.log" Oct 08 16:51:15 crc kubenswrapper[4945]: I1008 16:51:15.502655 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d5fb75f8-44p9s_bac08d54-03b7-4701-b9c7-69cc37c2eb3d/placement-log/0.log" Oct 08 16:51:15 crc kubenswrapper[4945]: I1008 16:51:15.659077 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/init-config-reloader/0.log" Oct 08 16:51:15 crc kubenswrapper[4945]: I1008 16:51:15.852056 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/config-reloader/0.log" Oct 08 16:51:15 crc kubenswrapper[4945]: I1008 16:51:15.861600 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/init-config-reloader/0.log" Oct 08 16:51:15 crc kubenswrapper[4945]: I1008 16:51:15.910240 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/prometheus/0.log" Oct 08 16:51:16 crc kubenswrapper[4945]: I1008 16:51:16.048526 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/thanos-sidecar/0.log" Oct 08 16:51:16 crc kubenswrapper[4945]: I1008 16:51:16.118425 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a252dc40-e13b-400d-a2b4-8b7e99d9eca1/setup-container/0.log" Oct 08 16:51:16 crc kubenswrapper[4945]: I1008 16:51:16.373540 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a252dc40-e13b-400d-a2b4-8b7e99d9eca1/setup-container/0.log" Oct 08 16:51:16 crc kubenswrapper[4945]: I1008 16:51:16.491882 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a252dc40-e13b-400d-a2b4-8b7e99d9eca1/rabbitmq/0.log" Oct 08 16:51:16 crc kubenswrapper[4945]: I1008 16:51:16.612813 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_420a6f35-1c79-416f-a7f4-8b9f46e337bc/setup-container/0.log" Oct 08 16:51:16 crc kubenswrapper[4945]: I1008 16:51:16.901018 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_420a6f35-1c79-416f-a7f4-8b9f46e337bc/rabbitmq/0.log" Oct 08 16:51:16 crc kubenswrapper[4945]: I1008 16:51:16.910431 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_420a6f35-1c79-416f-a7f4-8b9f46e337bc/setup-container/0.log" Oct 08 16:51:17 crc kubenswrapper[4945]: I1008 16:51:17.152337 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e51e86c4-2107-417e-9a3d-383d7a60da48/setup-container/0.log" Oct 08 16:51:17 crc kubenswrapper[4945]: I1008 16:51:17.423993 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e51e86c4-2107-417e-9a3d-383d7a60da48/setup-container/0.log" Oct 08 16:51:17 crc kubenswrapper[4945]: I1008 16:51:17.462494 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e51e86c4-2107-417e-9a3d-383d7a60da48/rabbitmq/0.log" Oct 08 16:51:17 crc kubenswrapper[4945]: I1008 16:51:17.628136 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k_47e40524-5bcd-4e95-9353-c17f7c23476f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:17 crc kubenswrapper[4945]: I1008 16:51:17.651791 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2sgnl_9afefec4-1f33-4022-9f2d-58f7e48e90f8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:17 crc kubenswrapper[4945]: I1008 16:51:17.898916 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8_5acda85d-4b0e-4468-b19e-f2a13dc16935/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:18 crc kubenswrapper[4945]: I1008 16:51:18.104999 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dfrnn_3a5d1441-c9ee-46f1-8004-361e72448f64/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:18 crc kubenswrapper[4945]: I1008 16:51:18.237807 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-bzn58_02a8cc7d-9ec0-4840-9023-f2550c396fcc/ssh-known-hosts-edpm-deployment/0.log" Oct 08 16:51:18 crc kubenswrapper[4945]: I1008 16:51:18.513697 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7cfddd55fc-6xk6v_2e5af0dd-db9f-40a7-941c-1c08c0bc5be4/proxy-server/0.log" Oct 08 16:51:18 crc kubenswrapper[4945]: I1008 16:51:18.736575 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7cfddd55fc-6xk6v_2e5af0dd-db9f-40a7-941c-1c08c0bc5be4/proxy-httpd/0.log" Oct 08 16:51:18 crc kubenswrapper[4945]: I1008 16:51:18.953892 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nqt9s_77c71453-5d71-4c06-b1d8-06561a1b2778/swift-ring-rebalance/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.083756 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-auditor/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.207369 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-reaper/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.322495 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-replicator/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.364965 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-server/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.438174 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-auditor/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.575821 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-server/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.596034 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-replicator/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.704385 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-updater/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.832531 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-auditor/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.885304 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-expirer/0.log" Oct 08 16:51:19 crc kubenswrapper[4945]: I1008 16:51:19.948191 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-replicator/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.072897 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-server/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.219866 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/rsync/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.224500 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-updater/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.279377 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/swift-recon-cron/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.492656 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v89t5_d7a976cb-af44-4232-8415-da43e89bff8a/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.686040 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_af253961-462b-4f6a-95dc-088795589466/test-operator-logs-container/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.727992 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0f3aa285-da71-4df6-8188-cc528008a61b/tempest-tests-tempest-tests-runner/0.log" Oct 08 16:51:20 crc kubenswrapper[4945]: I1008 16:51:20.984568 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2_53ec9f64-c34e-4765-b4ef-fd44c75f003e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:51:22 crc kubenswrapper[4945]: I1008 16:51:22.146424 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_099759e8-5f3f-463b-9be1-5c6ee209ce0a/watcher-applier/0.log" Oct 08 16:51:22 crc kubenswrapper[4945]: I1008 16:51:22.224579 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435/watcher-api-log/0.log" Oct 08 16:51:25 crc kubenswrapper[4945]: I1008 16:51:25.024732 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:51:25 crc kubenswrapper[4945]: E1008 16:51:25.025533 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:51:26 crc kubenswrapper[4945]: I1008 16:51:26.033919 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_ba695d96-8b81-4753-bd49-4686716797d7/watcher-decision-engine/0.log" Oct 08 16:51:26 crc kubenswrapper[4945]: I1008 16:51:26.518913 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435/watcher-api/0.log" Oct 08 16:51:36 crc kubenswrapper[4945]: I1008 16:51:36.422733 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ceb183ce-9008-4a1a-8b6b-26d47b2b246d/memcached/0.log" Oct 08 16:51:38 crc kubenswrapper[4945]: I1008 16:51:38.024331 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:51:38 crc kubenswrapper[4945]: E1008 16:51:38.024811 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:51:52 crc kubenswrapper[4945]: I1008 16:51:52.030628 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:51:52 crc kubenswrapper[4945]: E1008 16:51:52.032535 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:52:06 crc kubenswrapper[4945]: I1008 16:52:06.025152 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:52:06 crc kubenswrapper[4945]: E1008 16:52:06.026012 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:52:17 crc kubenswrapper[4945]: I1008 16:52:17.246214 4945 generic.go:334] "Generic (PLEG): container finished" podID="f8742445-70b3-474f-b4d0-ce61044ba8c7" containerID="4c9b3a7be6c353ced2412ef28494ce7b75f4b2b57673fd9351fd891b6a539bbb" exitCode=0 Oct 08 16:52:17 crc kubenswrapper[4945]: I1008 16:52:17.246258 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" event={"ID":"f8742445-70b3-474f-b4d0-ce61044ba8c7","Type":"ContainerDied","Data":"4c9b3a7be6c353ced2412ef28494ce7b75f4b2b57673fd9351fd891b6a539bbb"} Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.383436 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.416853 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-jnm5r"] Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.426977 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-jnm5r"] Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.522448 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8742445-70b3-474f-b4d0-ce61044ba8c7-host\") pod \"f8742445-70b3-474f-b4d0-ce61044ba8c7\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.522544 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8742445-70b3-474f-b4d0-ce61044ba8c7-host" (OuterVolumeSpecName: "host") pod "f8742445-70b3-474f-b4d0-ce61044ba8c7" (UID: "f8742445-70b3-474f-b4d0-ce61044ba8c7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.522573 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrg62\" (UniqueName: \"kubernetes.io/projected/f8742445-70b3-474f-b4d0-ce61044ba8c7-kube-api-access-lrg62\") pod \"f8742445-70b3-474f-b4d0-ce61044ba8c7\" (UID: \"f8742445-70b3-474f-b4d0-ce61044ba8c7\") " Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.523259 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8742445-70b3-474f-b4d0-ce61044ba8c7-host\") on node \"crc\" DevicePath \"\"" Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.528487 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8742445-70b3-474f-b4d0-ce61044ba8c7-kube-api-access-lrg62" (OuterVolumeSpecName: "kube-api-access-lrg62") pod "f8742445-70b3-474f-b4d0-ce61044ba8c7" (UID: "f8742445-70b3-474f-b4d0-ce61044ba8c7"). InnerVolumeSpecName "kube-api-access-lrg62". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:52:18 crc kubenswrapper[4945]: I1008 16:52:18.625212 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrg62\" (UniqueName: \"kubernetes.io/projected/f8742445-70b3-474f-b4d0-ce61044ba8c7-kube-api-access-lrg62\") on node \"crc\" DevicePath \"\"" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.269427 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a810958e4b7f9061858d29b514e759f904edb1eded12d4191083f43a34a42ac" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.269495 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-jnm5r" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.591351 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-6wdgk"] Oct 08 16:52:19 crc kubenswrapper[4945]: E1008 16:52:19.591876 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8742445-70b3-474f-b4d0-ce61044ba8c7" containerName="container-00" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.591893 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8742445-70b3-474f-b4d0-ce61044ba8c7" containerName="container-00" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.592201 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8742445-70b3-474f-b4d0-ce61044ba8c7" containerName="container-00" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.592950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.746518 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/359f305c-d4a9-4bfd-871c-62591bb982f7-host\") pod \"crc-debug-6wdgk\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.746578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blngm\" (UniqueName: \"kubernetes.io/projected/359f305c-d4a9-4bfd-871c-62591bb982f7-kube-api-access-blngm\") pod \"crc-debug-6wdgk\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.849657 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/359f305c-d4a9-4bfd-871c-62591bb982f7-host\") pod \"crc-debug-6wdgk\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.849957 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blngm\" (UniqueName: \"kubernetes.io/projected/359f305c-d4a9-4bfd-871c-62591bb982f7-kube-api-access-blngm\") pod \"crc-debug-6wdgk\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.849815 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/359f305c-d4a9-4bfd-871c-62591bb982f7-host\") pod \"crc-debug-6wdgk\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.877039 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blngm\" (UniqueName: \"kubernetes.io/projected/359f305c-d4a9-4bfd-871c-62591bb982f7-kube-api-access-blngm\") pod \"crc-debug-6wdgk\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:19 crc kubenswrapper[4945]: I1008 16:52:19.913521 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:20 crc kubenswrapper[4945]: I1008 16:52:20.034882 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8742445-70b3-474f-b4d0-ce61044ba8c7" path="/var/lib/kubelet/pods/f8742445-70b3-474f-b4d0-ce61044ba8c7/volumes" Oct 08 16:52:20 crc kubenswrapper[4945]: I1008 16:52:20.281317 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" event={"ID":"359f305c-d4a9-4bfd-871c-62591bb982f7","Type":"ContainerStarted","Data":"8ad3dba6dc18b9c13199a70ea3016b23c4b8dfec6a3740309dc5421279cd64a8"} Oct 08 16:52:20 crc kubenswrapper[4945]: I1008 16:52:20.281527 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" event={"ID":"359f305c-d4a9-4bfd-871c-62591bb982f7","Type":"ContainerStarted","Data":"b1d9dccd0be43681d7be78b26b6d39ff8e81629bb3ccac6a078a261e729080a4"} Oct 08 16:52:20 crc kubenswrapper[4945]: I1008 16:52:20.303938 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" podStartSLOduration=1.3039190330000001 podStartE2EDuration="1.303919033s" podCreationTimestamp="2025-10-08 16:52:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 16:52:20.292462057 +0000 UTC m=+5949.646376968" watchObservedRunningTime="2025-10-08 16:52:20.303919033 +0000 UTC m=+5949.657833934" Oct 08 16:52:21 crc kubenswrapper[4945]: I1008 16:52:21.024412 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:52:21 crc kubenswrapper[4945]: E1008 16:52:21.024761 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:52:21 crc kubenswrapper[4945]: I1008 16:52:21.289948 4945 generic.go:334] "Generic (PLEG): container finished" podID="359f305c-d4a9-4bfd-871c-62591bb982f7" containerID="8ad3dba6dc18b9c13199a70ea3016b23c4b8dfec6a3740309dc5421279cd64a8" exitCode=0 Oct 08 16:52:21 crc kubenswrapper[4945]: I1008 16:52:21.289994 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" event={"ID":"359f305c-d4a9-4bfd-871c-62591bb982f7","Type":"ContainerDied","Data":"8ad3dba6dc18b9c13199a70ea3016b23c4b8dfec6a3740309dc5421279cd64a8"} Oct 08 16:52:22 crc kubenswrapper[4945]: I1008 16:52:22.410502 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:22 crc kubenswrapper[4945]: I1008 16:52:22.592269 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/359f305c-d4a9-4bfd-871c-62591bb982f7-host\") pod \"359f305c-d4a9-4bfd-871c-62591bb982f7\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " Oct 08 16:52:22 crc kubenswrapper[4945]: I1008 16:52:22.592383 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/359f305c-d4a9-4bfd-871c-62591bb982f7-host" (OuterVolumeSpecName: "host") pod "359f305c-d4a9-4bfd-871c-62591bb982f7" (UID: "359f305c-d4a9-4bfd-871c-62591bb982f7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 16:52:22 crc kubenswrapper[4945]: I1008 16:52:22.592433 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blngm\" (UniqueName: \"kubernetes.io/projected/359f305c-d4a9-4bfd-871c-62591bb982f7-kube-api-access-blngm\") pod \"359f305c-d4a9-4bfd-871c-62591bb982f7\" (UID: \"359f305c-d4a9-4bfd-871c-62591bb982f7\") " Oct 08 16:52:22 crc kubenswrapper[4945]: I1008 16:52:22.592948 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/359f305c-d4a9-4bfd-871c-62591bb982f7-host\") on node \"crc\" DevicePath \"\"" Oct 08 16:52:22 crc kubenswrapper[4945]: I1008 16:52:22.604122 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/359f305c-d4a9-4bfd-871c-62591bb982f7-kube-api-access-blngm" (OuterVolumeSpecName: "kube-api-access-blngm") pod "359f305c-d4a9-4bfd-871c-62591bb982f7" (UID: "359f305c-d4a9-4bfd-871c-62591bb982f7"). InnerVolumeSpecName "kube-api-access-blngm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:52:22 crc kubenswrapper[4945]: I1008 16:52:22.694333 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blngm\" (UniqueName: \"kubernetes.io/projected/359f305c-d4a9-4bfd-871c-62591bb982f7-kube-api-access-blngm\") on node \"crc\" DevicePath \"\"" Oct 08 16:52:23 crc kubenswrapper[4945]: I1008 16:52:23.316403 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" event={"ID":"359f305c-d4a9-4bfd-871c-62591bb982f7","Type":"ContainerDied","Data":"b1d9dccd0be43681d7be78b26b6d39ff8e81629bb3ccac6a078a261e729080a4"} Oct 08 16:52:23 crc kubenswrapper[4945]: I1008 16:52:23.316856 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1d9dccd0be43681d7be78b26b6d39ff8e81629bb3ccac6a078a261e729080a4" Oct 08 16:52:23 crc kubenswrapper[4945]: I1008 16:52:23.316644 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-6wdgk" Oct 08 16:52:29 crc kubenswrapper[4945]: I1008 16:52:29.983998 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-6wdgk"] Oct 08 16:52:29 crc kubenswrapper[4945]: I1008 16:52:29.991920 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-6wdgk"] Oct 08 16:52:30 crc kubenswrapper[4945]: I1008 16:52:30.040432 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="359f305c-d4a9-4bfd-871c-62591bb982f7" path="/var/lib/kubelet/pods/359f305c-d4a9-4bfd-871c-62591bb982f7/volumes" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.141710 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-d97zg"] Oct 08 16:52:31 crc kubenswrapper[4945]: E1008 16:52:31.142403 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="359f305c-d4a9-4bfd-871c-62591bb982f7" containerName="container-00" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.142416 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="359f305c-d4a9-4bfd-871c-62591bb982f7" containerName="container-00" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.142633 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="359f305c-d4a9-4bfd-871c-62591bb982f7" containerName="container-00" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.143382 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.172572 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjtgf\" (UniqueName: \"kubernetes.io/projected/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-kube-api-access-jjtgf\") pod \"crc-debug-d97zg\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.172643 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-host\") pod \"crc-debug-d97zg\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.274936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjtgf\" (UniqueName: \"kubernetes.io/projected/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-kube-api-access-jjtgf\") pod \"crc-debug-d97zg\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.275016 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-host\") pod \"crc-debug-d97zg\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.275161 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-host\") pod \"crc-debug-d97zg\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.294703 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjtgf\" (UniqueName: \"kubernetes.io/projected/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-kube-api-access-jjtgf\") pod \"crc-debug-d97zg\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:31 crc kubenswrapper[4945]: I1008 16:52:31.472993 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:32 crc kubenswrapper[4945]: I1008 16:52:32.398628 4945 generic.go:334] "Generic (PLEG): container finished" podID="e5db8d63-e81c-4cf2-b348-3b93b4bb29ce" containerID="b521721e0981d7b0f3e8c5def1c618cc6d0b2451e16927f3632c7ad0d5218d08" exitCode=0 Oct 08 16:52:32 crc kubenswrapper[4945]: I1008 16:52:32.398719 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-d97zg" event={"ID":"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce","Type":"ContainerDied","Data":"b521721e0981d7b0f3e8c5def1c618cc6d0b2451e16927f3632c7ad0d5218d08"} Oct 08 16:52:32 crc kubenswrapper[4945]: I1008 16:52:32.399026 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/crc-debug-d97zg" event={"ID":"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce","Type":"ContainerStarted","Data":"ada2482e8a2c92a60a2c912c6fef43702a0156407db1bbaf4e79c26638c75c02"} Oct 08 16:52:32 crc kubenswrapper[4945]: I1008 16:52:32.481658 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-d97zg"] Oct 08 16:52:32 crc kubenswrapper[4945]: I1008 16:52:32.490799 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lzfw/crc-debug-d97zg"] Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.023876 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:52:33 crc kubenswrapper[4945]: E1008 16:52:33.024133 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.540941 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.725140 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-host\") pod \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.725288 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-host" (OuterVolumeSpecName: "host") pod "e5db8d63-e81c-4cf2-b348-3b93b4bb29ce" (UID: "e5db8d63-e81c-4cf2-b348-3b93b4bb29ce"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.725961 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjtgf\" (UniqueName: \"kubernetes.io/projected/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-kube-api-access-jjtgf\") pod \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\" (UID: \"e5db8d63-e81c-4cf2-b348-3b93b4bb29ce\") " Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.726475 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-host\") on node \"crc\" DevicePath \"\"" Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.746371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-kube-api-access-jjtgf" (OuterVolumeSpecName: "kube-api-access-jjtgf") pod "e5db8d63-e81c-4cf2-b348-3b93b4bb29ce" (UID: "e5db8d63-e81c-4cf2-b348-3b93b4bb29ce"). InnerVolumeSpecName "kube-api-access-jjtgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.828485 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjtgf\" (UniqueName: \"kubernetes.io/projected/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce-kube-api-access-jjtgf\") on node \"crc\" DevicePath \"\"" Oct 08 16:52:33 crc kubenswrapper[4945]: I1008 16:52:33.915757 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/util/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.036620 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5db8d63-e81c-4cf2-b348-3b93b4bb29ce" path="/var/lib/kubelet/pods/e5db8d63-e81c-4cf2-b348-3b93b4bb29ce/volumes" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.136229 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/pull/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.153384 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/util/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.177009 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/pull/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.349675 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/util/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.350483 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/pull/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.372591 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/extract/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.423688 4945 scope.go:117] "RemoveContainer" containerID="b521721e0981d7b0f3e8c5def1c618cc6d0b2451e16927f3632c7ad0d5218d08" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.423719 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/crc-debug-d97zg" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.510697 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-7b8pm_2747410b-55fa-4a7f-9146-5662ee48cce7/kube-rbac-proxy/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.623448 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-h27l8_441d050d-f359-46d8-8c29-0b1506c1f240/kube-rbac-proxy/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.641677 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-7b8pm_2747410b-55fa-4a7f-9146-5662ee48cce7/manager/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.737857 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-h27l8_441d050d-f359-46d8-8c29-0b1506c1f240/manager/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.830916 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-4fjrf_c3ae7aef-2f32-4d2d-b8be-e29dd91ede71/kube-rbac-proxy/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.862248 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-4fjrf_c3ae7aef-2f32-4d2d-b8be-e29dd91ede71/manager/0.log" Oct 08 16:52:34 crc kubenswrapper[4945]: I1008 16:52:34.966251 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-kx5k7_5c314cb3-27af-4a08-b96b-91f9a70f7524/kube-rbac-proxy/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.040339 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-kx5k7_5c314cb3-27af-4a08-b96b-91f9a70f7524/manager/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.150634 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-62hpj_84979368-496d-43c1-b792-20c167a2e5d5/kube-rbac-proxy/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.243039 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-62hpj_84979368-496d-43c1-b792-20c167a2e5d5/manager/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.291170 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-j9vft_b32201a1-e841-4c4c-8988-0b866243b255/kube-rbac-proxy/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.348287 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-j9vft_b32201a1-e841-4c4c-8988-0b866243b255/manager/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.464456 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-r9mp8_f11222c0-9d0f-4d3e-945f-b991135e4235/kube-rbac-proxy/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.648197 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-hk5rr_25cc3669-bdfd-44eb-83d0-85ca303250ea/kube-rbac-proxy/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.700795 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-hk5rr_25cc3669-bdfd-44eb-83d0-85ca303250ea/manager/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.748359 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-r9mp8_f11222c0-9d0f-4d3e-945f-b991135e4235/manager/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.857783 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-lr6hn_344902f1-cf0c-4390-8f2c-2fe41cf9ccd9/kube-rbac-proxy/0.log" Oct 08 16:52:35 crc kubenswrapper[4945]: I1008 16:52:35.969217 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-lr6hn_344902f1-cf0c-4390-8f2c-2fe41cf9ccd9/manager/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.034307 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ht7cm_cbe3a47b-033a-410d-b2be-bd38e4b9c0ab/kube-rbac-proxy/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.036970 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ht7cm_cbe3a47b-033a-410d-b2be-bd38e4b9c0ab/manager/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.128337 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-p7bfn_b2ca2442-7e23-40a8-9ca1-1ef11251d34d/kube-rbac-proxy/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.224844 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-p7bfn_b2ca2442-7e23-40a8-9ca1-1ef11251d34d/manager/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.334301 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-t56f2_1112a081-2d10-48e2-b345-8ef841c5dc50/kube-rbac-proxy/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.396340 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-t56f2_1112a081-2d10-48e2-b345-8ef841c5dc50/manager/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.430903 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-lwppj_3501f2c3-1b0d-4f7a-9fe2-2516f0088993/kube-rbac-proxy/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.587500 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-lwppj_3501f2c3-1b0d-4f7a-9fe2-2516f0088993/manager/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.640818 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-tb4rv_9555aa6e-7b8f-4354-96e8-e639719e1b56/kube-rbac-proxy/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.683246 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-tb4rv_9555aa6e-7b8f-4354-96e8-e639719e1b56/manager/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.790051 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt_a8d00f8c-42c6-4d99-bb76-bdd491597e3a/kube-rbac-proxy/0.log" Oct 08 16:52:36 crc kubenswrapper[4945]: I1008 16:52:36.794890 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt_a8d00f8c-42c6-4d99-bb76-bdd491597e3a/manager/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.018385 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56688bc7d6-h2wfx_c42bd8a4-d1ed-446c-9295-03cfe33588de/kube-rbac-proxy/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.116541 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75967d884c-qslk5_ef574f9e-15f9-42fe-8593-5575c31fe0d9/kube-rbac-proxy/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.326385 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-65hg6_9844bfac-ebec-4692-abda-eacfe73d5b3a/registry-server/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.417756 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75967d884c-qslk5_ef574f9e-15f9-42fe-8593-5575c31fe0d9/operator/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.527361 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-5t8fx_400ef757-79e4-4dd6-8c0a-2440835ac048/kube-rbac-proxy/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.711332 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-5t8fx_400ef757-79e4-4dd6-8c0a-2440835ac048/manager/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.843176 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-c8h57_efd39ce1-ac6e-449d-b32f-fc451ffde97e/kube-rbac-proxy/0.log" Oct 08 16:52:37 crc kubenswrapper[4945]: I1008 16:52:37.948773 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-c8h57_efd39ce1-ac6e-449d-b32f-fc451ffde97e/manager/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.058184 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx_1d068bb6-a523-4903-95b7-7c62862c2bb3/operator/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.176760 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qxbqr_43c420bf-cf21-4caa-90ce-0a9009d3bc45/kube-rbac-proxy/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.242375 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qxbqr_43c420bf-cf21-4caa-90ce-0a9009d3bc45/manager/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.384239 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56688bc7d6-h2wfx_c42bd8a4-d1ed-446c-9295-03cfe33588de/manager/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.385964 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-rn55t_f8b50124-bf7a-497b-bdc6-42bbbe564a7d/kube-rbac-proxy/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.490654 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-dwfj9_28b9d7cf-ad46-4c2a-a942-f7b20c054216/kube-rbac-proxy/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.607986 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-dwfj9_28b9d7cf-ad46-4c2a-a942-f7b20c054216/manager/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.639223 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-rn55t_f8b50124-bf7a-497b-bdc6-42bbbe564a7d/manager/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.684760 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-df6684d58-22z48_e0437b00-8fcc-412a-90fa-32ea56c71421/kube-rbac-proxy/0.log" Oct 08 16:52:38 crc kubenswrapper[4945]: I1008 16:52:38.760439 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-df6684d58-22z48_e0437b00-8fcc-412a-90fa-32ea56c71421/manager/0.log" Oct 08 16:52:46 crc kubenswrapper[4945]: I1008 16:52:46.024751 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:52:46 crc kubenswrapper[4945]: E1008 16:52:46.025760 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:52:53 crc kubenswrapper[4945]: I1008 16:52:53.243949 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mn52j_f973b44d-6296-408b-9d14-31d56fdd1021/control-plane-machine-set-operator/0.log" Oct 08 16:52:53 crc kubenswrapper[4945]: I1008 16:52:53.365093 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j72zk_ec002292-f388-497c-b114-fb7818730f4d/kube-rbac-proxy/0.log" Oct 08 16:52:53 crc kubenswrapper[4945]: I1008 16:52:53.437100 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j72zk_ec002292-f388-497c-b114-fb7818730f4d/machine-api-operator/0.log" Oct 08 16:52:57 crc kubenswrapper[4945]: I1008 16:52:57.028337 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:52:57 crc kubenswrapper[4945]: E1008 16:52:57.030248 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:53:04 crc kubenswrapper[4945]: I1008 16:53:04.476048 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xhl5p_8a063d76-5a7a-49f7-b19c-fac4fd303a90/cert-manager-controller/0.log" Oct 08 16:53:04 crc kubenswrapper[4945]: I1008 16:53:04.633816 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-2pms8_a020faff-9bc7-4f4c-83a4-9e21f352e5f2/cert-manager-cainjector/0.log" Oct 08 16:53:04 crc kubenswrapper[4945]: I1008 16:53:04.647851 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-q9846_ec83ad94-ee5b-4984-8381-bda1289c4ef8/cert-manager-webhook/0.log" Oct 08 16:53:09 crc kubenswrapper[4945]: I1008 16:53:09.025033 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:53:09 crc kubenswrapper[4945]: E1008 16:53:09.026026 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:53:15 crc kubenswrapper[4945]: I1008 16:53:15.726765 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-7zpln_d4d9183c-cff3-4bed-82e0-41a654969322/nmstate-console-plugin/0.log" Oct 08 16:53:15 crc kubenswrapper[4945]: I1008 16:53:15.879621 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6q4xl_835b6fb9-28f9-4c64-8049-05dd404be8cc/nmstate-handler/0.log" Oct 08 16:53:15 crc kubenswrapper[4945]: I1008 16:53:15.921611 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-kfqmx_0afd8411-7a96-44b9-b1af-82a05419fa24/kube-rbac-proxy/0.log" Oct 08 16:53:15 crc kubenswrapper[4945]: I1008 16:53:15.942438 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-kfqmx_0afd8411-7a96-44b9-b1af-82a05419fa24/nmstate-metrics/0.log" Oct 08 16:53:16 crc kubenswrapper[4945]: I1008 16:53:16.095194 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-z75g2_9d7e5f1a-5f1b-4291-8dee-de24611c5ff7/nmstate-operator/0.log" Oct 08 16:53:16 crc kubenswrapper[4945]: I1008 16:53:16.166215 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-59hrm_c84bb408-d873-4b03-a07b-dfb913fdc64b/nmstate-webhook/0.log" Oct 08 16:53:21 crc kubenswrapper[4945]: I1008 16:53:21.025308 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:53:21 crc kubenswrapper[4945]: E1008 16:53:21.026358 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:53:28 crc kubenswrapper[4945]: I1008 16:53:28.637765 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-nnbmf_919b5756-8417-47d1-acd7-fbfd2c01d246/kube-rbac-proxy/0.log" Oct 08 16:53:28 crc kubenswrapper[4945]: I1008 16:53:28.860354 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-nnbmf_919b5756-8417-47d1-acd7-fbfd2c01d246/controller/0.log" Oct 08 16:53:28 crc kubenswrapper[4945]: I1008 16:53:28.961462 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.080033 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.080185 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.122957 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.180241 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.540833 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.541020 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.560847 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.634772 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.751968 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.756664 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.763460 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.863301 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/controller/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.975396 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/kube-rbac-proxy/0.log" Oct 08 16:53:29 crc kubenswrapper[4945]: I1008 16:53:29.991868 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/frr-metrics/0.log" Oct 08 16:53:30 crc kubenswrapper[4945]: I1008 16:53:30.091281 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/kube-rbac-proxy-frr/0.log" Oct 08 16:53:30 crc kubenswrapper[4945]: I1008 16:53:30.261755 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/reloader/0.log" Oct 08 16:53:30 crc kubenswrapper[4945]: I1008 16:53:30.337523 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-h5s5z_1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c/frr-k8s-webhook-server/0.log" Oct 08 16:53:30 crc kubenswrapper[4945]: I1008 16:53:30.609320 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-575bbc8666-4t674_13c15125-a872-459e-ac9a-d37356f47418/manager/0.log" Oct 08 16:53:30 crc kubenswrapper[4945]: I1008 16:53:30.703749 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d999db554-m5t6x_132bc709-76d1-41fc-ae20-04141e80a6ac/webhook-server/0.log" Oct 08 16:53:30 crc kubenswrapper[4945]: I1008 16:53:30.899176 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dftb8_283c80b9-bfe4-4c26-852d-9ba6d14009d4/kube-rbac-proxy/0.log" Oct 08 16:53:31 crc kubenswrapper[4945]: I1008 16:53:31.472036 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dftb8_283c80b9-bfe4-4c26-852d-9ba6d14009d4/speaker/0.log" Oct 08 16:53:31 crc kubenswrapper[4945]: I1008 16:53:31.749538 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/frr/0.log" Oct 08 16:53:34 crc kubenswrapper[4945]: I1008 16:53:34.024338 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:53:34 crc kubenswrapper[4945]: E1008 16:53:34.024874 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:53:42 crc kubenswrapper[4945]: I1008 16:53:42.813352 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/util/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.016209 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/pull/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.016336 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/pull/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.083822 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/util/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.122427 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/util/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.224390 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/extract/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.228898 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/pull/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.344339 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/util/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.572600 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/util/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.663784 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/util/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.683917 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/pull/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.684326 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/pull/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.722953 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/pull/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.842715 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/extract/0.log" Oct 08 16:53:43 crc kubenswrapper[4945]: I1008 16:53:43.882582 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-utilities/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.052859 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-content/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.056616 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-content/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.108772 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-utilities/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.299452 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-utilities/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.313717 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-content/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.513315 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-utilities/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.689536 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-content/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.711976 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-content/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.733846 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-utilities/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.908303 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-content/0.log" Oct 08 16:53:44 crc kubenswrapper[4945]: I1008 16:53:44.953683 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-utilities/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.317885 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/util/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.380802 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/registry-server/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.541179 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/pull/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.566783 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/util/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.654863 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/pull/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.770270 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/util/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.806439 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/pull/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.899886 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/extract/0.log" Oct 08 16:53:45 crc kubenswrapper[4945]: I1008 16:53:45.986950 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-vnm62_e99374e4-2a46-4044-881c-9bdb32f4a215/marketplace-operator/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.182700 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-utilities/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.367462 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-utilities/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.373627 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-content/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.376783 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/registry-server/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.386008 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-content/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.559389 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-utilities/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.564845 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-content/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.606918 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-utilities/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.756387 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-utilities/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.769458 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-content/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.795610 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-content/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.966687 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-content/0.log" Oct 08 16:53:46 crc kubenswrapper[4945]: I1008 16:53:46.977143 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-utilities/0.log" Oct 08 16:53:47 crc kubenswrapper[4945]: I1008 16:53:47.024660 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:53:47 crc kubenswrapper[4945]: E1008 16:53:47.024913 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 16:53:47 crc kubenswrapper[4945]: I1008 16:53:47.751555 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/registry-server/0.log" Oct 08 16:53:48 crc kubenswrapper[4945]: I1008 16:53:48.446077 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/registry-server/0.log" Oct 08 16:53:58 crc kubenswrapper[4945]: I1008 16:53:58.024596 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:53:58 crc kubenswrapper[4945]: I1008 16:53:58.230880 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-g4w8k_18169f3c-d245-4b63-8ea7-d6d3bd70cce4/prometheus-operator/0.log" Oct 08 16:53:58 crc kubenswrapper[4945]: I1008 16:53:58.345697 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp_658dd3b7-2709-4785-aa43-64f041d46b7a/prometheus-operator-admission-webhook/0.log" Oct 08 16:53:58 crc kubenswrapper[4945]: I1008 16:53:58.384219 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd_fed6f952-f438-4a3b-bd8c-f2832a1c6e8d/prometheus-operator-admission-webhook/0.log" Oct 08 16:53:58 crc kubenswrapper[4945]: I1008 16:53:58.617448 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-sw8w6_d39d5733-b343-4a33-866b-220a8191eda9/perses-operator/0.log" Oct 08 16:53:58 crc kubenswrapper[4945]: I1008 16:53:58.638706 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-qrx7m_09a37473-34c8-43ff-a839-3f712f3d0605/operator/0.log" Oct 08 16:53:59 crc kubenswrapper[4945]: I1008 16:53:59.232284 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"237f1f106bec3e491f7495488dc33c96187402c9198bafca4b532ff6e25ad8c3"} Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.672424 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dtrrd"] Oct 08 16:55:35 crc kubenswrapper[4945]: E1008 16:55:35.673374 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5db8d63-e81c-4cf2-b348-3b93b4bb29ce" containerName="container-00" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.673389 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5db8d63-e81c-4cf2-b348-3b93b4bb29ce" containerName="container-00" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.673647 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5db8d63-e81c-4cf2-b348-3b93b4bb29ce" containerName="container-00" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.675462 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.687494 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtrrd"] Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.759101 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-catalog-content\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.759599 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrg6w\" (UniqueName: \"kubernetes.io/projected/30970bae-4f65-424b-9772-381c0416ede7-kube-api-access-mrg6w\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.759905 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-utilities\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.861640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-catalog-content\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.861809 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrg6w\" (UniqueName: \"kubernetes.io/projected/30970bae-4f65-424b-9772-381c0416ede7-kube-api-access-mrg6w\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.861913 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-utilities\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.862197 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-catalog-content\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.862432 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-utilities\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.881782 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrg6w\" (UniqueName: \"kubernetes.io/projected/30970bae-4f65-424b-9772-381c0416ede7-kube-api-access-mrg6w\") pod \"certified-operators-dtrrd\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:35 crc kubenswrapper[4945]: I1008 16:55:35.997055 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:36 crc kubenswrapper[4945]: I1008 16:55:36.878344 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtrrd"] Oct 08 16:55:37 crc kubenswrapper[4945]: I1008 16:55:37.238094 4945 generic.go:334] "Generic (PLEG): container finished" podID="30970bae-4f65-424b-9772-381c0416ede7" containerID="7cf0601b85dfb1c02179992c956a5cde53f50598dcd2f378d05198513c41bc68" exitCode=0 Oct 08 16:55:37 crc kubenswrapper[4945]: I1008 16:55:37.238215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtrrd" event={"ID":"30970bae-4f65-424b-9772-381c0416ede7","Type":"ContainerDied","Data":"7cf0601b85dfb1c02179992c956a5cde53f50598dcd2f378d05198513c41bc68"} Oct 08 16:55:37 crc kubenswrapper[4945]: I1008 16:55:37.238390 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtrrd" event={"ID":"30970bae-4f65-424b-9772-381c0416ede7","Type":"ContainerStarted","Data":"d3ba3e77eb2dcd0de08dd432d7886841999f196fa63b731eed232244c72f44c6"} Oct 08 16:55:37 crc kubenswrapper[4945]: I1008 16:55:37.240759 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 16:55:39 crc kubenswrapper[4945]: I1008 16:55:39.258959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtrrd" event={"ID":"30970bae-4f65-424b-9772-381c0416ede7","Type":"ContainerStarted","Data":"26806adb181535cda6f446ddcc14d98ef1970105203456deed8cb8b1bb3515ae"} Oct 08 16:55:42 crc kubenswrapper[4945]: I1008 16:55:42.288830 4945 generic.go:334] "Generic (PLEG): container finished" podID="30970bae-4f65-424b-9772-381c0416ede7" containerID="26806adb181535cda6f446ddcc14d98ef1970105203456deed8cb8b1bb3515ae" exitCode=0 Oct 08 16:55:42 crc kubenswrapper[4945]: I1008 16:55:42.288914 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtrrd" event={"ID":"30970bae-4f65-424b-9772-381c0416ede7","Type":"ContainerDied","Data":"26806adb181535cda6f446ddcc14d98ef1970105203456deed8cb8b1bb3515ae"} Oct 08 16:55:46 crc kubenswrapper[4945]: I1008 16:55:46.337327 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtrrd" event={"ID":"30970bae-4f65-424b-9772-381c0416ede7","Type":"ContainerStarted","Data":"0a0f404e0f5ffc775fbb6ff6e788f300dfd15b48664e9f764f96d3d212d6c111"} Oct 08 16:55:46 crc kubenswrapper[4945]: I1008 16:55:46.363763 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dtrrd" podStartSLOduration=3.4073126990000002 podStartE2EDuration="11.363740948s" podCreationTimestamp="2025-10-08 16:55:35 +0000 UTC" firstStartedPulling="2025-10-08 16:55:37.240568561 +0000 UTC m=+6146.594483462" lastFinishedPulling="2025-10-08 16:55:45.19699681 +0000 UTC m=+6154.550911711" observedRunningTime="2025-10-08 16:55:46.353293127 +0000 UTC m=+6155.707208038" watchObservedRunningTime="2025-10-08 16:55:46.363740948 +0000 UTC m=+6155.717655849" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.071395 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h8c2k"] Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.074712 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.148777 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8c2k"] Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.161607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-utilities\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.161782 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pmpq\" (UniqueName: \"kubernetes.io/projected/af151e2c-853d-48d2-9207-7e84a2ebcccf-kube-api-access-6pmpq\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.161856 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-catalog-content\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.263312 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-utilities\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.263438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pmpq\" (UniqueName: \"kubernetes.io/projected/af151e2c-853d-48d2-9207-7e84a2ebcccf-kube-api-access-6pmpq\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.263483 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-catalog-content\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.263878 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-utilities\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.263968 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-catalog-content\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.285323 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pmpq\" (UniqueName: \"kubernetes.io/projected/af151e2c-853d-48d2-9207-7e84a2ebcccf-kube-api-access-6pmpq\") pod \"redhat-operators-h8c2k\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.395880 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:55:54 crc kubenswrapper[4945]: I1008 16:55:54.912823 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8c2k"] Oct 08 16:55:55 crc kubenswrapper[4945]: I1008 16:55:55.430373 4945 generic.go:334] "Generic (PLEG): container finished" podID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerID="c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3" exitCode=0 Oct 08 16:55:55 crc kubenswrapper[4945]: I1008 16:55:55.430641 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8c2k" event={"ID":"af151e2c-853d-48d2-9207-7e84a2ebcccf","Type":"ContainerDied","Data":"c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3"} Oct 08 16:55:55 crc kubenswrapper[4945]: I1008 16:55:55.430667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8c2k" event={"ID":"af151e2c-853d-48d2-9207-7e84a2ebcccf","Type":"ContainerStarted","Data":"dec71df6f876615456423e34f4624ba818336282a0b064e28ae9d396029f2b11"} Oct 08 16:55:55 crc kubenswrapper[4945]: I1008 16:55:55.997441 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:55 crc kubenswrapper[4945]: I1008 16:55:55.998031 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:56 crc kubenswrapper[4945]: I1008 16:55:56.071450 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:56 crc kubenswrapper[4945]: I1008 16:55:56.503899 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:55:57 crc kubenswrapper[4945]: I1008 16:55:57.482906 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8c2k" event={"ID":"af151e2c-853d-48d2-9207-7e84a2ebcccf","Type":"ContainerStarted","Data":"349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879"} Oct 08 16:55:58 crc kubenswrapper[4945]: I1008 16:55:58.449727 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtrrd"] Oct 08 16:55:59 crc kubenswrapper[4945]: I1008 16:55:59.505699 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dtrrd" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="registry-server" containerID="cri-o://0a0f404e0f5ffc775fbb6ff6e788f300dfd15b48664e9f764f96d3d212d6c111" gracePeriod=2 Oct 08 16:56:00 crc kubenswrapper[4945]: I1008 16:56:00.517162 4945 generic.go:334] "Generic (PLEG): container finished" podID="30970bae-4f65-424b-9772-381c0416ede7" containerID="0a0f404e0f5ffc775fbb6ff6e788f300dfd15b48664e9f764f96d3d212d6c111" exitCode=0 Oct 08 16:56:00 crc kubenswrapper[4945]: I1008 16:56:00.517242 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtrrd" event={"ID":"30970bae-4f65-424b-9772-381c0416ede7","Type":"ContainerDied","Data":"0a0f404e0f5ffc775fbb6ff6e788f300dfd15b48664e9f764f96d3d212d6c111"} Oct 08 16:56:00 crc kubenswrapper[4945]: I1008 16:56:00.942541 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.118403 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-utilities\") pod \"30970bae-4f65-424b-9772-381c0416ede7\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.118527 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-catalog-content\") pod \"30970bae-4f65-424b-9772-381c0416ede7\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.118636 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrg6w\" (UniqueName: \"kubernetes.io/projected/30970bae-4f65-424b-9772-381c0416ede7-kube-api-access-mrg6w\") pod \"30970bae-4f65-424b-9772-381c0416ede7\" (UID: \"30970bae-4f65-424b-9772-381c0416ede7\") " Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.119360 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-utilities" (OuterVolumeSpecName: "utilities") pod "30970bae-4f65-424b-9772-381c0416ede7" (UID: "30970bae-4f65-424b-9772-381c0416ede7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.119599 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.125652 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30970bae-4f65-424b-9772-381c0416ede7-kube-api-access-mrg6w" (OuterVolumeSpecName: "kube-api-access-mrg6w") pod "30970bae-4f65-424b-9772-381c0416ede7" (UID: "30970bae-4f65-424b-9772-381c0416ede7"). InnerVolumeSpecName "kube-api-access-mrg6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.174993 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30970bae-4f65-424b-9772-381c0416ede7" (UID: "30970bae-4f65-424b-9772-381c0416ede7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.221162 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30970bae-4f65-424b-9772-381c0416ede7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.221197 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrg6w\" (UniqueName: \"kubernetes.io/projected/30970bae-4f65-424b-9772-381c0416ede7-kube-api-access-mrg6w\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.529770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtrrd" event={"ID":"30970bae-4f65-424b-9772-381c0416ede7","Type":"ContainerDied","Data":"d3ba3e77eb2dcd0de08dd432d7886841999f196fa63b731eed232244c72f44c6"} Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.530147 4945 scope.go:117] "RemoveContainer" containerID="0a0f404e0f5ffc775fbb6ff6e788f300dfd15b48664e9f764f96d3d212d6c111" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.529933 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtrrd" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.561317 4945 scope.go:117] "RemoveContainer" containerID="26806adb181535cda6f446ddcc14d98ef1970105203456deed8cb8b1bb3515ae" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.588208 4945 scope.go:117] "RemoveContainer" containerID="7cf0601b85dfb1c02179992c956a5cde53f50598dcd2f378d05198513c41bc68" Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.599054 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtrrd"] Oct 08 16:56:01 crc kubenswrapper[4945]: I1008 16:56:01.609829 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dtrrd"] Oct 08 16:56:02 crc kubenswrapper[4945]: I1008 16:56:02.042391 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30970bae-4f65-424b-9772-381c0416ede7" path="/var/lib/kubelet/pods/30970bae-4f65-424b-9772-381c0416ede7/volumes" Oct 08 16:56:13 crc kubenswrapper[4945]: I1008 16:56:13.676498 4945 generic.go:334] "Generic (PLEG): container finished" podID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerID="349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879" exitCode=0 Oct 08 16:56:13 crc kubenswrapper[4945]: I1008 16:56:13.676560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8c2k" event={"ID":"af151e2c-853d-48d2-9207-7e84a2ebcccf","Type":"ContainerDied","Data":"349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879"} Oct 08 16:56:15 crc kubenswrapper[4945]: I1008 16:56:15.715737 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8c2k" event={"ID":"af151e2c-853d-48d2-9207-7e84a2ebcccf","Type":"ContainerStarted","Data":"f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd"} Oct 08 16:56:15 crc kubenswrapper[4945]: I1008 16:56:15.743592 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h8c2k" podStartSLOduration=2.6365290850000003 podStartE2EDuration="21.743570128s" podCreationTimestamp="2025-10-08 16:55:54 +0000 UTC" firstStartedPulling="2025-10-08 16:55:55.432801424 +0000 UTC m=+6164.786716325" lastFinishedPulling="2025-10-08 16:56:14.539842457 +0000 UTC m=+6183.893757368" observedRunningTime="2025-10-08 16:56:15.73722516 +0000 UTC m=+6185.091140081" watchObservedRunningTime="2025-10-08 16:56:15.743570128 +0000 UTC m=+6185.097485029" Oct 08 16:56:19 crc kubenswrapper[4945]: I1008 16:56:19.184402 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:56:19 crc kubenswrapper[4945]: I1008 16:56:19.184933 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:56:24 crc kubenswrapper[4945]: I1008 16:56:24.396495 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:56:24 crc kubenswrapper[4945]: I1008 16:56:24.398253 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:56:24 crc kubenswrapper[4945]: I1008 16:56:24.449575 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:56:24 crc kubenswrapper[4945]: I1008 16:56:24.838124 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:56:25 crc kubenswrapper[4945]: I1008 16:56:25.265268 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h8c2k"] Oct 08 16:56:25 crc kubenswrapper[4945]: I1008 16:56:25.810261 4945 generic.go:334] "Generic (PLEG): container finished" podID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerID="989abd9d6a3b49c6c2c43ddc5c4a6e032d727076a4aa8187fc703d10c4878b91" exitCode=0 Oct 08 16:56:25 crc kubenswrapper[4945]: I1008 16:56:25.810411 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" event={"ID":"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5","Type":"ContainerDied","Data":"989abd9d6a3b49c6c2c43ddc5c4a6e032d727076a4aa8187fc703d10c4878b91"} Oct 08 16:56:25 crc kubenswrapper[4945]: I1008 16:56:25.812686 4945 scope.go:117] "RemoveContainer" containerID="989abd9d6a3b49c6c2c43ddc5c4a6e032d727076a4aa8187fc703d10c4878b91" Oct 08 16:56:26 crc kubenswrapper[4945]: I1008 16:56:26.383580 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7lzfw_must-gather-j8dz7_ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5/gather/0.log" Oct 08 16:56:26 crc kubenswrapper[4945]: I1008 16:56:26.824789 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h8c2k" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="registry-server" containerID="cri-o://f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd" gracePeriod=2 Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.312285 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.493060 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-utilities\") pod \"af151e2c-853d-48d2-9207-7e84a2ebcccf\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.493124 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pmpq\" (UniqueName: \"kubernetes.io/projected/af151e2c-853d-48d2-9207-7e84a2ebcccf-kube-api-access-6pmpq\") pod \"af151e2c-853d-48d2-9207-7e84a2ebcccf\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.493162 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-catalog-content\") pod \"af151e2c-853d-48d2-9207-7e84a2ebcccf\" (UID: \"af151e2c-853d-48d2-9207-7e84a2ebcccf\") " Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.494241 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-utilities" (OuterVolumeSpecName: "utilities") pod "af151e2c-853d-48d2-9207-7e84a2ebcccf" (UID: "af151e2c-853d-48d2-9207-7e84a2ebcccf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.509990 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af151e2c-853d-48d2-9207-7e84a2ebcccf-kube-api-access-6pmpq" (OuterVolumeSpecName: "kube-api-access-6pmpq") pod "af151e2c-853d-48d2-9207-7e84a2ebcccf" (UID: "af151e2c-853d-48d2-9207-7e84a2ebcccf"). InnerVolumeSpecName "kube-api-access-6pmpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.590088 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af151e2c-853d-48d2-9207-7e84a2ebcccf" (UID: "af151e2c-853d-48d2-9207-7e84a2ebcccf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.595660 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.595976 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pmpq\" (UniqueName: \"kubernetes.io/projected/af151e2c-853d-48d2-9207-7e84a2ebcccf-kube-api-access-6pmpq\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.596134 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af151e2c-853d-48d2-9207-7e84a2ebcccf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.839756 4945 generic.go:334] "Generic (PLEG): container finished" podID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerID="f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd" exitCode=0 Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.839801 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8c2k" event={"ID":"af151e2c-853d-48d2-9207-7e84a2ebcccf","Type":"ContainerDied","Data":"f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd"} Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.839825 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8c2k" event={"ID":"af151e2c-853d-48d2-9207-7e84a2ebcccf","Type":"ContainerDied","Data":"dec71df6f876615456423e34f4624ba818336282a0b064e28ae9d396029f2b11"} Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.839844 4945 scope.go:117] "RemoveContainer" containerID="f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.840126 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8c2k" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.868144 4945 scope.go:117] "RemoveContainer" containerID="349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.886059 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h8c2k"] Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.897378 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h8c2k"] Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.910556 4945 scope.go:117] "RemoveContainer" containerID="c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.970359 4945 scope.go:117] "RemoveContainer" containerID="f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd" Oct 08 16:56:27 crc kubenswrapper[4945]: E1008 16:56:27.970784 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd\": container with ID starting with f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd not found: ID does not exist" containerID="f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.970853 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd"} err="failed to get container status \"f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd\": rpc error: code = NotFound desc = could not find container \"f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd\": container with ID starting with f790812aca40f70236b895aa857b89ab5a64e5fa0a79255c764ca84af3f8fdcd not found: ID does not exist" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.970885 4945 scope.go:117] "RemoveContainer" containerID="349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879" Oct 08 16:56:27 crc kubenswrapper[4945]: E1008 16:56:27.971346 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879\": container with ID starting with 349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879 not found: ID does not exist" containerID="349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.971394 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879"} err="failed to get container status \"349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879\": rpc error: code = NotFound desc = could not find container \"349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879\": container with ID starting with 349105bd49ee1ed6a170665370c2179d2f41368d69db2499ed2276eab0f63879 not found: ID does not exist" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.971414 4945 scope.go:117] "RemoveContainer" containerID="c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3" Oct 08 16:56:27 crc kubenswrapper[4945]: E1008 16:56:27.971656 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3\": container with ID starting with c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3 not found: ID does not exist" containerID="c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3" Oct 08 16:56:27 crc kubenswrapper[4945]: I1008 16:56:27.971679 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3"} err="failed to get container status \"c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3\": rpc error: code = NotFound desc = could not find container \"c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3\": container with ID starting with c4238aac56f152d0772da15bc0cd97135013686050ce47684420bc83681adbe3 not found: ID does not exist" Oct 08 16:56:28 crc kubenswrapper[4945]: I1008 16:56:28.035647 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" path="/var/lib/kubelet/pods/af151e2c-853d-48d2-9207-7e84a2ebcccf/volumes" Oct 08 16:56:36 crc kubenswrapper[4945]: I1008 16:56:36.633562 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lzfw/must-gather-j8dz7"] Oct 08 16:56:36 crc kubenswrapper[4945]: I1008 16:56:36.634183 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerName="copy" containerID="cri-o://120f101da0c28389cb8d9753e1475870d26f0e0233bff15618ea04dd01c79ace" gracePeriod=2 Oct 08 16:56:36 crc kubenswrapper[4945]: I1008 16:56:36.640162 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lzfw/must-gather-j8dz7"] Oct 08 16:56:36 crc kubenswrapper[4945]: I1008 16:56:36.941396 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7lzfw_must-gather-j8dz7_ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5/copy/0.log" Oct 08 16:56:36 crc kubenswrapper[4945]: I1008 16:56:36.943371 4945 generic.go:334] "Generic (PLEG): container finished" podID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerID="120f101da0c28389cb8d9753e1475870d26f0e0233bff15618ea04dd01c79ace" exitCode=143 Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.138816 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7lzfw_must-gather-j8dz7_ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5/copy/0.log" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.139564 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.296696 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l4nm\" (UniqueName: \"kubernetes.io/projected/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-kube-api-access-9l4nm\") pod \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.296796 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-must-gather-output\") pod \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\" (UID: \"ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5\") " Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.304135 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-kube-api-access-9l4nm" (OuterVolumeSpecName: "kube-api-access-9l4nm") pod "ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" (UID: "ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5"). InnerVolumeSpecName "kube-api-access-9l4nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.399282 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l4nm\" (UniqueName: \"kubernetes.io/projected/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-kube-api-access-9l4nm\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.506840 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" (UID: "ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.604100 4945 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.953437 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7lzfw_must-gather-j8dz7_ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5/copy/0.log" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.953796 4945 scope.go:117] "RemoveContainer" containerID="120f101da0c28389cb8d9753e1475870d26f0e0233bff15618ea04dd01c79ace" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.953892 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lzfw/must-gather-j8dz7" Oct 08 16:56:37 crc kubenswrapper[4945]: I1008 16:56:37.977932 4945 scope.go:117] "RemoveContainer" containerID="989abd9d6a3b49c6c2c43ddc5c4a6e032d727076a4aa8187fc703d10c4878b91" Oct 08 16:56:38 crc kubenswrapper[4945]: I1008 16:56:38.038075 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" path="/var/lib/kubelet/pods/ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5/volumes" Oct 08 16:56:47 crc kubenswrapper[4945]: I1008 16:56:47.409543 4945 scope.go:117] "RemoveContainer" containerID="4c9b3a7be6c353ced2412ef28494ce7b75f4b2b57673fd9351fd891b6a539bbb" Oct 08 16:56:49 crc kubenswrapper[4945]: I1008 16:56:49.185254 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:56:49 crc kubenswrapper[4945]: I1008 16:56:49.185638 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.296209 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wf46x/must-gather-klcdw"] Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298043 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="extract-content" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.298143 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="extract-content" Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298212 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="extract-utilities" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.298271 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="extract-utilities" Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298340 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerName="gather" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.298399 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerName="gather" Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298463 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="extract-content" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.298521 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="extract-content" Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298580 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="extract-utilities" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.298632 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="extract-utilities" Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298703 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="registry-server" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.298757 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="registry-server" Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298824 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerName="copy" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.298926 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerName="copy" Oct 08 16:57:04 crc kubenswrapper[4945]: E1008 16:57:04.298989 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="registry-server" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.299050 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="registry-server" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.299336 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="af151e2c-853d-48d2-9207-7e84a2ebcccf" containerName="registry-server" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.299424 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerName="gather" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.299558 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7b6960-2f9c-49cc-bbc4-af23dfcaf0a5" containerName="copy" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.299624 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="30970bae-4f65-424b-9772-381c0416ede7" containerName="registry-server" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.300781 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.302598 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wf46x"/"default-dockercfg-77qml" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.312444 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wf46x/must-gather-klcdw"] Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.314632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efe6cd45-99e7-42ec-8acd-c42cc681a319-must-gather-output\") pod \"must-gather-klcdw\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.314845 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vn7d\" (UniqueName: \"kubernetes.io/projected/efe6cd45-99e7-42ec-8acd-c42cc681a319-kube-api-access-2vn7d\") pod \"must-gather-klcdw\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.320506 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wf46x"/"openshift-service-ca.crt" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.320561 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wf46x"/"kube-root-ca.crt" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.416989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efe6cd45-99e7-42ec-8acd-c42cc681a319-must-gather-output\") pod \"must-gather-klcdw\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.417359 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vn7d\" (UniqueName: \"kubernetes.io/projected/efe6cd45-99e7-42ec-8acd-c42cc681a319-kube-api-access-2vn7d\") pod \"must-gather-klcdw\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.417490 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efe6cd45-99e7-42ec-8acd-c42cc681a319-must-gather-output\") pod \"must-gather-klcdw\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.435050 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vn7d\" (UniqueName: \"kubernetes.io/projected/efe6cd45-99e7-42ec-8acd-c42cc681a319-kube-api-access-2vn7d\") pod \"must-gather-klcdw\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:04 crc kubenswrapper[4945]: I1008 16:57:04.618186 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 16:57:05 crc kubenswrapper[4945]: I1008 16:57:05.158943 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wf46x/must-gather-klcdw"] Oct 08 16:57:05 crc kubenswrapper[4945]: I1008 16:57:05.258209 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/must-gather-klcdw" event={"ID":"efe6cd45-99e7-42ec-8acd-c42cc681a319","Type":"ContainerStarted","Data":"c6d8016b7dc26f33fc7574ffc15cda42a262fcf6c8f3b756ba72da7f57f8b1e2"} Oct 08 16:57:06 crc kubenswrapper[4945]: I1008 16:57:06.268734 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/must-gather-klcdw" event={"ID":"efe6cd45-99e7-42ec-8acd-c42cc681a319","Type":"ContainerStarted","Data":"dc8b0657328ec206321bc98de7304ea09705e8be01a083932d93cb11c8f29cd0"} Oct 08 16:57:06 crc kubenswrapper[4945]: I1008 16:57:06.269043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/must-gather-klcdw" event={"ID":"efe6cd45-99e7-42ec-8acd-c42cc681a319","Type":"ContainerStarted","Data":"c1d6d83a6ef1940a5cc25840cb95ffb6265cb5b51d050d2a72e83cc47eefbb1a"} Oct 08 16:57:06 crc kubenswrapper[4945]: I1008 16:57:06.286211 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wf46x/must-gather-klcdw" podStartSLOduration=2.286187565 podStartE2EDuration="2.286187565s" podCreationTimestamp="2025-10-08 16:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 16:57:06.283313313 +0000 UTC m=+6235.637228204" watchObservedRunningTime="2025-10-08 16:57:06.286187565 +0000 UTC m=+6235.640102466" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.324282 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wf46x/crc-debug-gfrtq"] Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.326165 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.456758 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-host\") pod \"crc-debug-gfrtq\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.457369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2v6p\" (UniqueName: \"kubernetes.io/projected/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-kube-api-access-q2v6p\") pod \"crc-debug-gfrtq\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.559857 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-host\") pod \"crc-debug-gfrtq\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.559981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2v6p\" (UniqueName: \"kubernetes.io/projected/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-kube-api-access-q2v6p\") pod \"crc-debug-gfrtq\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.560382 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-host\") pod \"crc-debug-gfrtq\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.596467 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2v6p\" (UniqueName: \"kubernetes.io/projected/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-kube-api-access-q2v6p\") pod \"crc-debug-gfrtq\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: I1008 16:57:09.651452 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:57:09 crc kubenswrapper[4945]: W1008 16:57:09.718475 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fedd4a7_1a9e_4a15_a766_f54cbc78b33f.slice/crio-05ad78ffd3e4fb8c6256e96a422d1a9f692db3a911a819e008e5938f9f146d2d WatchSource:0}: Error finding container 05ad78ffd3e4fb8c6256e96a422d1a9f692db3a911a819e008e5938f9f146d2d: Status 404 returned error can't find the container with id 05ad78ffd3e4fb8c6256e96a422d1a9f692db3a911a819e008e5938f9f146d2d Oct 08 16:57:10 crc kubenswrapper[4945]: I1008 16:57:10.304509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" event={"ID":"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f","Type":"ContainerStarted","Data":"914c38fd8a3473fb366ef1b84dcabd2e44c1dd378d9e0e7020bcf767b5d981c4"} Oct 08 16:57:10 crc kubenswrapper[4945]: I1008 16:57:10.305264 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" event={"ID":"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f","Type":"ContainerStarted","Data":"05ad78ffd3e4fb8c6256e96a422d1a9f692db3a911a819e008e5938f9f146d2d"} Oct 08 16:57:10 crc kubenswrapper[4945]: I1008 16:57:10.324319 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" podStartSLOduration=1.324304416 podStartE2EDuration="1.324304416s" podCreationTimestamp="2025-10-08 16:57:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 16:57:10.315943797 +0000 UTC m=+6239.669858708" watchObservedRunningTime="2025-10-08 16:57:10.324304416 +0000 UTC m=+6239.678219317" Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.184375 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.185035 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.185090 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.186026 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"237f1f106bec3e491f7495488dc33c96187402c9198bafca4b532ff6e25ad8c3"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.186101 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://237f1f106bec3e491f7495488dc33c96187402c9198bafca4b532ff6e25ad8c3" gracePeriod=600 Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.399217 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="237f1f106bec3e491f7495488dc33c96187402c9198bafca4b532ff6e25ad8c3" exitCode=0 Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.399312 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"237f1f106bec3e491f7495488dc33c96187402c9198bafca4b532ff6e25ad8c3"} Oct 08 16:57:19 crc kubenswrapper[4945]: I1008 16:57:19.399685 4945 scope.go:117] "RemoveContainer" containerID="bdc25205c3abb5f57508571ee96ca8b3cafa5c6f1dab6effb536762a74a4b9f1" Oct 08 16:57:20 crc kubenswrapper[4945]: I1008 16:57:20.411937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6"} Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.537673 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5qt48"] Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.541365 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.551015 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qt48"] Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.647707 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-catalog-content\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.647943 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzfpw\" (UniqueName: \"kubernetes.io/projected/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-kube-api-access-gzfpw\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.647985 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-utilities\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.749875 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzfpw\" (UniqueName: \"kubernetes.io/projected/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-kube-api-access-gzfpw\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.750317 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-utilities\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.750631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-catalog-content\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.751130 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-utilities\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.751214 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-catalog-content\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.778719 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzfpw\" (UniqueName: \"kubernetes.io/projected/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-kube-api-access-gzfpw\") pod \"redhat-marketplace-5qt48\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:22 crc kubenswrapper[4945]: I1008 16:58:22.899145 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:23 crc kubenswrapper[4945]: I1008 16:58:23.402464 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qt48"] Oct 08 16:58:24 crc kubenswrapper[4945]: I1008 16:58:24.056390 4945 generic.go:334] "Generic (PLEG): container finished" podID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerID="5dfcc59e1941c295beaece3f8fe55a34a9e53fa0014574ef172573b7744fa59b" exitCode=0 Oct 08 16:58:24 crc kubenswrapper[4945]: I1008 16:58:24.056561 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qt48" event={"ID":"4bbc9ac2-3300-431b-a23d-96accb2ed9bf","Type":"ContainerDied","Data":"5dfcc59e1941c295beaece3f8fe55a34a9e53fa0014574ef172573b7744fa59b"} Oct 08 16:58:24 crc kubenswrapper[4945]: I1008 16:58:24.056723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qt48" event={"ID":"4bbc9ac2-3300-431b-a23d-96accb2ed9bf","Type":"ContainerStarted","Data":"aba369bc1d270a507788c38be6f5aae1384e2d4226b171e182900e1d4ed177b1"} Oct 08 16:58:26 crc kubenswrapper[4945]: I1008 16:58:26.075012 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qt48" event={"ID":"4bbc9ac2-3300-431b-a23d-96accb2ed9bf","Type":"ContainerStarted","Data":"8a370dfffee38412d90f07e316688ee8cca70476067c7b70482f17161891d954"} Oct 08 16:58:27 crc kubenswrapper[4945]: I1008 16:58:27.088730 4945 generic.go:334] "Generic (PLEG): container finished" podID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerID="8a370dfffee38412d90f07e316688ee8cca70476067c7b70482f17161891d954" exitCode=0 Oct 08 16:58:27 crc kubenswrapper[4945]: I1008 16:58:27.088836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qt48" event={"ID":"4bbc9ac2-3300-431b-a23d-96accb2ed9bf","Type":"ContainerDied","Data":"8a370dfffee38412d90f07e316688ee8cca70476067c7b70482f17161891d954"} Oct 08 16:58:28 crc kubenswrapper[4945]: I1008 16:58:28.111777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qt48" event={"ID":"4bbc9ac2-3300-431b-a23d-96accb2ed9bf","Type":"ContainerStarted","Data":"5778243015e0a845389adbdc3a204854c866635e35237b8e49ecd831e11f9d16"} Oct 08 16:58:28 crc kubenswrapper[4945]: I1008 16:58:28.197503 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5qt48" podStartSLOduration=2.404712466 podStartE2EDuration="6.197480446s" podCreationTimestamp="2025-10-08 16:58:22 +0000 UTC" firstStartedPulling="2025-10-08 16:58:24.059211793 +0000 UTC m=+6313.413126704" lastFinishedPulling="2025-10-08 16:58:27.851979733 +0000 UTC m=+6317.205894684" observedRunningTime="2025-10-08 16:58:28.140292712 +0000 UTC m=+6317.494207633" watchObservedRunningTime="2025-10-08 16:58:28.197480446 +0000 UTC m=+6317.551395347" Oct 08 16:58:32 crc kubenswrapper[4945]: I1008 16:58:32.899745 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:32 crc kubenswrapper[4945]: I1008 16:58:32.900635 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:32 crc kubenswrapper[4945]: I1008 16:58:32.947759 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:33 crc kubenswrapper[4945]: I1008 16:58:33.220022 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:33 crc kubenswrapper[4945]: I1008 16:58:33.263802 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qt48"] Oct 08 16:58:35 crc kubenswrapper[4945]: I1008 16:58:35.190888 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5qt48" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="registry-server" containerID="cri-o://5778243015e0a845389adbdc3a204854c866635e35237b8e49ecd831e11f9d16" gracePeriod=2 Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.206129 4945 generic.go:334] "Generic (PLEG): container finished" podID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerID="5778243015e0a845389adbdc3a204854c866635e35237b8e49ecd831e11f9d16" exitCode=0 Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.206507 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qt48" event={"ID":"4bbc9ac2-3300-431b-a23d-96accb2ed9bf","Type":"ContainerDied","Data":"5778243015e0a845389adbdc3a204854c866635e35237b8e49ecd831e11f9d16"} Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.659574 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.781381 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-catalog-content\") pod \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.781538 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzfpw\" (UniqueName: \"kubernetes.io/projected/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-kube-api-access-gzfpw\") pod \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.781611 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-utilities\") pod \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\" (UID: \"4bbc9ac2-3300-431b-a23d-96accb2ed9bf\") " Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.782637 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-utilities" (OuterVolumeSpecName: "utilities") pod "4bbc9ac2-3300-431b-a23d-96accb2ed9bf" (UID: "4bbc9ac2-3300-431b-a23d-96accb2ed9bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.790322 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-kube-api-access-gzfpw" (OuterVolumeSpecName: "kube-api-access-gzfpw") pod "4bbc9ac2-3300-431b-a23d-96accb2ed9bf" (UID: "4bbc9ac2-3300-431b-a23d-96accb2ed9bf"). InnerVolumeSpecName "kube-api-access-gzfpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.792331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bbc9ac2-3300-431b-a23d-96accb2ed9bf" (UID: "4bbc9ac2-3300-431b-a23d-96accb2ed9bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.795053 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5588d6b8d6-7nmw2_ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6/barbican-api/0.log" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.883594 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzfpw\" (UniqueName: \"kubernetes.io/projected/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-kube-api-access-gzfpw\") on node \"crc\" DevicePath \"\"" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.883629 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.883638 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbc9ac2-3300-431b-a23d-96accb2ed9bf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:58:36 crc kubenswrapper[4945]: I1008 16:58:36.898489 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5588d6b8d6-7nmw2_ebe5d0d3-f41d-4225-8b84-b8c16ee30ca6/barbican-api-log/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.147814 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b8d567b6d-zc6xh_66b89671-2c42-4e97-8a74-0e03bac0613a/barbican-keystone-listener/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.159150 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b8d567b6d-zc6xh_66b89671-2c42-4e97-8a74-0e03bac0613a/barbican-keystone-listener-log/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.217882 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qt48" event={"ID":"4bbc9ac2-3300-431b-a23d-96accb2ed9bf","Type":"ContainerDied","Data":"aba369bc1d270a507788c38be6f5aae1384e2d4226b171e182900e1d4ed177b1"} Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.217932 4945 scope.go:117] "RemoveContainer" containerID="5778243015e0a845389adbdc3a204854c866635e35237b8e49ecd831e11f9d16" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.218065 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qt48" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.250501 4945 scope.go:117] "RemoveContainer" containerID="8a370dfffee38412d90f07e316688ee8cca70476067c7b70482f17161891d954" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.275986 4945 scope.go:117] "RemoveContainer" containerID="5dfcc59e1941c295beaece3f8fe55a34a9e53fa0014574ef172573b7744fa59b" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.279500 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qt48"] Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.305177 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qt48"] Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.392606 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fc897d669-lcbrz_dfcd7c98-f494-4669-8e76-995ccb3e768b/barbican-worker-log/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.393689 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fc897d669-lcbrz_dfcd7c98-f494-4669-8e76-995ccb3e768b/barbican-worker/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.639965 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-m2vf2_de2e46b1-591e-4540-b69e-eee056133019/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.860070 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/proxy-httpd/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.891553 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/ceilometer-notification-agent/0.log" Oct 08 16:58:37 crc kubenswrapper[4945]: I1008 16:58:37.906086 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/ceilometer-central-agent/0.log" Oct 08 16:58:38 crc kubenswrapper[4945]: I1008 16:58:38.036424 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" path="/var/lib/kubelet/pods/4bbc9ac2-3300-431b-a23d-96accb2ed9bf/volumes" Oct 08 16:58:38 crc kubenswrapper[4945]: I1008 16:58:38.052607 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_950bbdee-1a42-4890-b678-abd7e6a65402/sg-core/0.log" Oct 08 16:58:38 crc kubenswrapper[4945]: I1008 16:58:38.304494 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c1bfcac1-b1e7-431f-b145-62a2da972214/cinder-api-log/0.log" Oct 08 16:58:38 crc kubenswrapper[4945]: I1008 16:58:38.710386 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_a2eda97e-ccd7-42e8-bca5-26402069c250/probe/0.log" Oct 08 16:58:38 crc kubenswrapper[4945]: I1008 16:58:38.784408 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c1bfcac1-b1e7-431f-b145-62a2da972214/cinder-api/0.log" Oct 08 16:58:38 crc kubenswrapper[4945]: I1008 16:58:38.893761 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_a2eda97e-ccd7-42e8-bca5-26402069c250/cinder-backup/0.log" Oct 08 16:58:39 crc kubenswrapper[4945]: I1008 16:58:39.058672 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_815a3456-3316-4348-a417-f274fda65c41/cinder-scheduler/0.log" Oct 08 16:58:39 crc kubenswrapper[4945]: I1008 16:58:39.146722 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_815a3456-3316-4348-a417-f274fda65c41/probe/0.log" Oct 08 16:58:39 crc kubenswrapper[4945]: I1008 16:58:39.461726 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_1c0d0627-7d92-4bd7-8f53-3cd9117aec3b/probe/0.log" Oct 08 16:58:39 crc kubenswrapper[4945]: I1008 16:58:39.463481 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_1c0d0627-7d92-4bd7-8f53-3cd9117aec3b/cinder-volume/0.log" Oct 08 16:58:39 crc kubenswrapper[4945]: I1008 16:58:39.697027 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e2f8bf86-ab2b-428a-9e89-2d71091927ef/probe/0.log" Oct 08 16:58:39 crc kubenswrapper[4945]: I1008 16:58:39.866126 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e2f8bf86-ab2b-428a-9e89-2d71091927ef/cinder-volume/0.log" Oct 08 16:58:39 crc kubenswrapper[4945]: I1008 16:58:39.954241 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-pnjkh_fcdb8897-6c75-48d4-bbdf-b4eb486331d5/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.145902 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-dzp4t_a6323567-dc30-45ed-8611-4925d983e0ef/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.192289 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tkrps_71b8f45e-f47d-432d-b22e-3fa39e8059dd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.318376 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f7b68cdfc-bhvjv_108d033e-a880-4b1d-ba80-293d36d0a58a/init/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.506586 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f7b68cdfc-bhvjv_108d033e-a880-4b1d-ba80-293d36d0a58a/init/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.574575 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-tkqp8_e59943cd-9572-4e6b-8a58-c0ec1283b2aa/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.737337 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f7b68cdfc-bhvjv_108d033e-a880-4b1d-ba80-293d36d0a58a/dnsmasq-dns/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.865848 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bc16ff2a-fe01-42bf-8b16-e4ba10315929/glance-httpd/0.log" Oct 08 16:58:40 crc kubenswrapper[4945]: I1008 16:58:40.869810 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bc16ff2a-fe01-42bf-8b16-e4ba10315929/glance-log/0.log" Oct 08 16:58:41 crc kubenswrapper[4945]: I1008 16:58:41.067141 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f2f0a0ac-d149-4fc1-878a-7e21188cd4ee/glance-httpd/0.log" Oct 08 16:58:41 crc kubenswrapper[4945]: I1008 16:58:41.117085 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f2f0a0ac-d149-4fc1-878a-7e21188cd4ee/glance-log/0.log" Oct 08 16:58:41 crc kubenswrapper[4945]: I1008 16:58:41.354589 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-29p9d_b8610fc4-8948-4247-91b3-78ca4e5e1767/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:41 crc kubenswrapper[4945]: I1008 16:58:41.363385 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c58b7c96-vjh9q_37d6d391-e363-42db-8a0e-4e837e4af4db/horizon/0.log" Oct 08 16:58:41 crc kubenswrapper[4945]: I1008 16:58:41.679492 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-twd5z_bad79416-d5f9-4e87-bab7-cbd0ea8047e9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:41 crc kubenswrapper[4945]: I1008 16:58:41.951678 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332321-qc7kq_ca3e1262-6ad2-4df4-b86d-dc17fc2339c6/keystone-cron/0.log" Oct 08 16:58:42 crc kubenswrapper[4945]: I1008 16:58:42.078632 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c58b7c96-vjh9q_37d6d391-e363-42db-8a0e-4e837e4af4db/horizon-log/0.log" Oct 08 16:58:42 crc kubenswrapper[4945]: I1008 16:58:42.141965 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_803b9d32-2d37-4fd2-96fd-b4001daf26ed/kube-state-metrics/0.log" Oct 08 16:58:42 crc kubenswrapper[4945]: I1008 16:58:42.408870 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-76mxw_8a77f15d-769f-493c-b410-eb13c9df8fe6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:42 crc kubenswrapper[4945]: I1008 16:58:42.446516 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-65b767dfc8-l5jkz_f748658a-398b-4b6e-9f27-d2a45e6f6827/keystone-api/0.log" Oct 08 16:58:42 crc kubenswrapper[4945]: I1008 16:58:42.960788 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5988468485-ctpxh_acfb96c2-ab3d-4e3f-86ce-c6befd371617/neutron-httpd/0.log" Oct 08 16:58:43 crc kubenswrapper[4945]: I1008 16:58:43.048996 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5988468485-ctpxh_acfb96c2-ab3d-4e3f-86ce-c6befd371617/neutron-api/0.log" Oct 08 16:58:43 crc kubenswrapper[4945]: I1008 16:58:43.180641 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-br94p_0c384ac4-a16d-4d25-b5e0-dc7eb33f7352/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:44 crc kubenswrapper[4945]: I1008 16:58:44.188412 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d1cf9a55-81f7-45e6-97b6-b05f4e7f07be/nova-cell0-conductor-conductor/0.log" Oct 08 16:58:44 crc kubenswrapper[4945]: I1008 16:58:44.972463 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ae15d968-e006-4d49-92aa-97cabf627769/nova-cell1-conductor-conductor/0.log" Oct 08 16:58:45 crc kubenswrapper[4945]: I1008 16:58:45.129411 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e8e508c6-7afc-4ae0-84bb-48dece116b1d/nova-api-log/0.log" Oct 08 16:58:45 crc kubenswrapper[4945]: I1008 16:58:45.526924 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e8e508c6-7afc-4ae0-84bb-48dece116b1d/nova-api-api/0.log" Oct 08 16:58:45 crc kubenswrapper[4945]: I1008 16:58:45.582336 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_7a3edb5a-ae0f-411f-b1f5-a9d591a831b6/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 16:58:45 crc kubenswrapper[4945]: I1008 16:58:45.708604 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-mpr75_8b121934-fc40-487d-91fe-cdc91272f2dd/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:45 crc kubenswrapper[4945]: I1008 16:58:45.870816 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1ce5750e-a063-42a1-bde2-0731599fa3bd/nova-metadata-log/0.log" Oct 08 16:58:46 crc kubenswrapper[4945]: I1008 16:58:46.519092 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1d5103f7-4fab-43a6-8989-1b505ac5d8d0/mysql-bootstrap/0.log" Oct 08 16:58:46 crc kubenswrapper[4945]: I1008 16:58:46.581582 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c17f16f2-5d8f-4178-a6b5-879444c7b262/nova-scheduler-scheduler/0.log" Oct 08 16:58:46 crc kubenswrapper[4945]: I1008 16:58:46.735715 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1d5103f7-4fab-43a6-8989-1b505ac5d8d0/mysql-bootstrap/0.log" Oct 08 16:58:46 crc kubenswrapper[4945]: I1008 16:58:46.833607 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1d5103f7-4fab-43a6-8989-1b505ac5d8d0/galera/0.log" Oct 08 16:58:47 crc kubenswrapper[4945]: I1008 16:58:47.114795 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_88f0eb1e-a2bc-4b2f-b14d-8cef0261334e/mysql-bootstrap/0.log" Oct 08 16:58:47 crc kubenswrapper[4945]: I1008 16:58:47.279802 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_88f0eb1e-a2bc-4b2f-b14d-8cef0261334e/mysql-bootstrap/0.log" Oct 08 16:58:47 crc kubenswrapper[4945]: I1008 16:58:47.323596 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_88f0eb1e-a2bc-4b2f-b14d-8cef0261334e/galera/0.log" Oct 08 16:58:47 crc kubenswrapper[4945]: I1008 16:58:47.522926 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_223e741f-c7cc-465f-a4bc-804bfff7a6ef/openstackclient/0.log" Oct 08 16:58:47 crc kubenswrapper[4945]: I1008 16:58:47.546831 4945 scope.go:117] "RemoveContainer" containerID="8ad3dba6dc18b9c13199a70ea3016b23c4b8dfec6a3740309dc5421279cd64a8" Oct 08 16:58:47 crc kubenswrapper[4945]: I1008 16:58:47.797590 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-cwwpc_1b70322d-ae60-46b4-828b-1f4a01cae6b6/ovn-controller/0.log" Oct 08 16:58:47 crc kubenswrapper[4945]: I1008 16:58:47.962335 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-x7dfs_2c0817bf-174f-4d4f-b155-041259384465/openstack-network-exporter/0.log" Oct 08 16:58:48 crc kubenswrapper[4945]: I1008 16:58:48.278312 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovsdb-server-init/0.log" Oct 08 16:58:48 crc kubenswrapper[4945]: I1008 16:58:48.519875 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovsdb-server-init/0.log" Oct 08 16:58:48 crc kubenswrapper[4945]: I1008 16:58:48.596630 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1ce5750e-a063-42a1-bde2-0731599fa3bd/nova-metadata-metadata/0.log" Oct 08 16:58:48 crc kubenswrapper[4945]: I1008 16:58:48.684004 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovsdb-server/0.log" Oct 08 16:58:48 crc kubenswrapper[4945]: I1008 16:58:48.868143 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-67zmn_7fa35d9f-678d-4820-9be9-7db07c5ffd07/ovs-vswitchd/0.log" Oct 08 16:58:48 crc kubenswrapper[4945]: I1008 16:58:48.938534 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-wv9ct_d32f281f-0e3f-4ac5-8463-7bb261bc557e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:49 crc kubenswrapper[4945]: I1008 16:58:49.091885 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f9854a42-6052-4425-a3a8-190bd5533f75/openstack-network-exporter/0.log" Oct 08 16:58:49 crc kubenswrapper[4945]: I1008 16:58:49.141419 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f9854a42-6052-4425-a3a8-190bd5533f75/ovn-northd/0.log" Oct 08 16:58:49 crc kubenswrapper[4945]: I1008 16:58:49.310838 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c4cb5094-ffbc-4fff-bbc4-0a3a35067782/openstack-network-exporter/0.log" Oct 08 16:58:49 crc kubenswrapper[4945]: I1008 16:58:49.370887 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c4cb5094-ffbc-4fff-bbc4-0a3a35067782/ovsdbserver-nb/0.log" Oct 08 16:58:49 crc kubenswrapper[4945]: I1008 16:58:49.759975 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec9d8fdb-5e82-4cb6-811b-053a28588453/ovsdbserver-sb/0.log" Oct 08 16:58:49 crc kubenswrapper[4945]: I1008 16:58:49.774487 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec9d8fdb-5e82-4cb6-811b-053a28588453/openstack-network-exporter/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.220978 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d5fb75f8-44p9s_bac08d54-03b7-4701-b9c7-69cc37c2eb3d/placement-api/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.255000 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d5fb75f8-44p9s_bac08d54-03b7-4701-b9c7-69cc37c2eb3d/placement-log/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.429190 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/init-config-reloader/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.616331 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/init-config-reloader/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.682149 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/config-reloader/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.688694 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/prometheus/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.946607 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ce8e6f8-1567-4bd4-8f2c-ada70838ec92/thanos-sidecar/0.log" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.992318 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2ns5j"] Oct 08 16:58:50 crc kubenswrapper[4945]: E1008 16:58:50.992744 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="extract-content" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.992760 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="extract-content" Oct 08 16:58:50 crc kubenswrapper[4945]: E1008 16:58:50.992782 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="registry-server" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.992789 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="registry-server" Oct 08 16:58:50 crc kubenswrapper[4945]: E1008 16:58:50.993340 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="extract-utilities" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.993355 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="extract-utilities" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.993636 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbc9ac2-3300-431b-a23d-96accb2ed9bf" containerName="registry-server" Oct 08 16:58:50 crc kubenswrapper[4945]: I1008 16:58:50.998173 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.024755 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2ns5j"] Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.083417 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dwvc\" (UniqueName: \"kubernetes.io/projected/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-kube-api-access-7dwvc\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.083522 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-utilities\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.083639 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-catalog-content\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.086681 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a252dc40-e13b-400d-a2b4-8b7e99d9eca1/setup-container/0.log" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.185166 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-catalog-content\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.185214 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dwvc\" (UniqueName: \"kubernetes.io/projected/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-kube-api-access-7dwvc\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.185317 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-utilities\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.186789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-catalog-content\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.187911 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-utilities\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.227072 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dwvc\" (UniqueName: \"kubernetes.io/projected/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-kube-api-access-7dwvc\") pod \"community-operators-2ns5j\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.295492 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a252dc40-e13b-400d-a2b4-8b7e99d9eca1/setup-container/0.log" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.332658 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.379184 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a252dc40-e13b-400d-a2b4-8b7e99d9eca1/rabbitmq/0.log" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.778966 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_420a6f35-1c79-416f-a7f4-8b9f46e337bc/setup-container/0.log" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.941543 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2ns5j"] Oct 08 16:58:51 crc kubenswrapper[4945]: W1008 16:58:51.942176 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c7b8ca1_c98f_48e2_bbaa_0914faec510a.slice/crio-0c7e897fe45e901f4f89f417b839f474bf804d81e0c2d3c3c2b3b7aca4e2c079 WatchSource:0}: Error finding container 0c7e897fe45e901f4f89f417b839f474bf804d81e0c2d3c3c2b3b7aca4e2c079: Status 404 returned error can't find the container with id 0c7e897fe45e901f4f89f417b839f474bf804d81e0c2d3c3c2b3b7aca4e2c079 Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.978685 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_420a6f35-1c79-416f-a7f4-8b9f46e337bc/setup-container/0.log" Oct 08 16:58:51 crc kubenswrapper[4945]: I1008 16:58:51.986042 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_420a6f35-1c79-416f-a7f4-8b9f46e337bc/rabbitmq/0.log" Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.225386 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e51e86c4-2107-417e-9a3d-383d7a60da48/setup-container/0.log" Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.368514 4945 generic.go:334] "Generic (PLEG): container finished" podID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerID="7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9" exitCode=0 Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.368555 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ns5j" event={"ID":"2c7b8ca1-c98f-48e2-bbaa-0914faec510a","Type":"ContainerDied","Data":"7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9"} Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.368580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ns5j" event={"ID":"2c7b8ca1-c98f-48e2-bbaa-0914faec510a","Type":"ContainerStarted","Data":"0c7e897fe45e901f4f89f417b839f474bf804d81e0c2d3c3c2b3b7aca4e2c079"} Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.503409 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e51e86c4-2107-417e-9a3d-383d7a60da48/setup-container/0.log" Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.566980 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e51e86c4-2107-417e-9a3d-383d7a60da48/rabbitmq/0.log" Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.815771 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-l5z2k_47e40524-5bcd-4e95-9353-c17f7c23476f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:52 crc kubenswrapper[4945]: I1008 16:58:52.835260 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2sgnl_9afefec4-1f33-4022-9f2d-58f7e48e90f8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:53 crc kubenswrapper[4945]: I1008 16:58:53.050173 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vwnw8_5acda85d-4b0e-4468-b19e-f2a13dc16935/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:53 crc kubenswrapper[4945]: I1008 16:58:53.289875 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dfrnn_3a5d1441-c9ee-46f1-8004-361e72448f64/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:53 crc kubenswrapper[4945]: I1008 16:58:53.402176 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-bzn58_02a8cc7d-9ec0-4840-9023-f2550c396fcc/ssh-known-hosts-edpm-deployment/0.log" Oct 08 16:58:53 crc kubenswrapper[4945]: I1008 16:58:53.749243 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7cfddd55fc-6xk6v_2e5af0dd-db9f-40a7-941c-1c08c0bc5be4/proxy-server/0.log" Oct 08 16:58:53 crc kubenswrapper[4945]: I1008 16:58:53.872546 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7cfddd55fc-6xk6v_2e5af0dd-db9f-40a7-941c-1c08c0bc5be4/proxy-httpd/0.log" Oct 08 16:58:53 crc kubenswrapper[4945]: I1008 16:58:53.953442 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nqt9s_77c71453-5d71-4c06-b1d8-06561a1b2778/swift-ring-rebalance/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.070947 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-auditor/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.149139 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-reaper/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.371240 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-server/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.386600 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-auditor/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.387158 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ns5j" event={"ID":"2c7b8ca1-c98f-48e2-bbaa-0914faec510a","Type":"ContainerStarted","Data":"e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e"} Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.394203 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/account-replicator/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.600714 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-server/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.601621 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-updater/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.670249 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/container-replicator/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.866595 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-auditor/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.933605 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-expirer/0.log" Oct 08 16:58:54 crc kubenswrapper[4945]: I1008 16:58:54.971425 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-replicator/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.125028 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-updater/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.145027 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/object-server/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.180653 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/rsync/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.318313 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e8011e91-9053-4a34-a58a-7d046d51536f/swift-recon-cron/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.474818 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v89t5_d7a976cb-af44-4232-8415-da43e89bff8a/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.573729 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0f3aa285-da71-4df6-8188-cc528008a61b/tempest-tests-tempest-tests-runner/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.808501 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_af253961-462b-4f6a-95dc-088795589466/test-operator-logs-container/0.log" Oct 08 16:58:55 crc kubenswrapper[4945]: I1008 16:58:55.959635 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-x2rr2_53ec9f64-c34e-4765-b4ef-fd44c75f003e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 16:58:57 crc kubenswrapper[4945]: I1008 16:58:57.343964 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_099759e8-5f3f-463b-9be1-5c6ee209ce0a/watcher-applier/0.log" Oct 08 16:58:58 crc kubenswrapper[4945]: I1008 16:58:58.236780 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435/watcher-api-log/0.log" Oct 08 16:58:59 crc kubenswrapper[4945]: I1008 16:58:59.457177 4945 generic.go:334] "Generic (PLEG): container finished" podID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerID="e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e" exitCode=0 Oct 08 16:58:59 crc kubenswrapper[4945]: I1008 16:58:59.457257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ns5j" event={"ID":"2c7b8ca1-c98f-48e2-bbaa-0914faec510a","Type":"ContainerDied","Data":"e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e"} Oct 08 16:59:01 crc kubenswrapper[4945]: I1008 16:59:01.376703 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_ba695d96-8b81-4753-bd49-4686716797d7/watcher-decision-engine/0.log" Oct 08 16:59:01 crc kubenswrapper[4945]: I1008 16:59:01.480447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ns5j" event={"ID":"2c7b8ca1-c98f-48e2-bbaa-0914faec510a","Type":"ContainerStarted","Data":"ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81"} Oct 08 16:59:01 crc kubenswrapper[4945]: I1008 16:59:01.499010 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2ns5j" podStartSLOduration=3.153003526 podStartE2EDuration="11.498994972s" podCreationTimestamp="2025-10-08 16:58:50 +0000 UTC" firstStartedPulling="2025-10-08 16:58:52.37019878 +0000 UTC m=+6341.724113681" lastFinishedPulling="2025-10-08 16:59:00.716190226 +0000 UTC m=+6350.070105127" observedRunningTime="2025-10-08 16:59:01.49774303 +0000 UTC m=+6350.851657931" watchObservedRunningTime="2025-10-08 16:59:01.498994972 +0000 UTC m=+6350.852909873" Oct 08 16:59:02 crc kubenswrapper[4945]: I1008 16:59:02.367082 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_f9d2ef74-dfaf-4dc0-8af8-4b7f6b262435/watcher-api/0.log" Oct 08 16:59:11 crc kubenswrapper[4945]: I1008 16:59:11.334203 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:59:11 crc kubenswrapper[4945]: I1008 16:59:11.334941 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:59:11 crc kubenswrapper[4945]: I1008 16:59:11.391923 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:59:11 crc kubenswrapper[4945]: I1008 16:59:11.634956 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:59:11 crc kubenswrapper[4945]: I1008 16:59:11.687332 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2ns5j"] Oct 08 16:59:13 crc kubenswrapper[4945]: I1008 16:59:13.593618 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2ns5j" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="registry-server" containerID="cri-o://ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81" gracePeriod=2 Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.069070 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.111225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-catalog-content\") pod \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.111338 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-utilities\") pod \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.111416 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dwvc\" (UniqueName: \"kubernetes.io/projected/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-kube-api-access-7dwvc\") pod \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\" (UID: \"2c7b8ca1-c98f-48e2-bbaa-0914faec510a\") " Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.117425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-utilities" (OuterVolumeSpecName: "utilities") pod "2c7b8ca1-c98f-48e2-bbaa-0914faec510a" (UID: "2c7b8ca1-c98f-48e2-bbaa-0914faec510a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.172254 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-kube-api-access-7dwvc" (OuterVolumeSpecName: "kube-api-access-7dwvc") pod "2c7b8ca1-c98f-48e2-bbaa-0914faec510a" (UID: "2c7b8ca1-c98f-48e2-bbaa-0914faec510a"). InnerVolumeSpecName "kube-api-access-7dwvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.208743 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c7b8ca1-c98f-48e2-bbaa-0914faec510a" (UID: "2c7b8ca1-c98f-48e2-bbaa-0914faec510a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.212960 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.212987 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.212998 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dwvc\" (UniqueName: \"kubernetes.io/projected/2c7b8ca1-c98f-48e2-bbaa-0914faec510a-kube-api-access-7dwvc\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.607293 4945 generic.go:334] "Generic (PLEG): container finished" podID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerID="ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81" exitCode=0 Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.607348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ns5j" event={"ID":"2c7b8ca1-c98f-48e2-bbaa-0914faec510a","Type":"ContainerDied","Data":"ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81"} Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.607384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ns5j" event={"ID":"2c7b8ca1-c98f-48e2-bbaa-0914faec510a","Type":"ContainerDied","Data":"0c7e897fe45e901f4f89f417b839f474bf804d81e0c2d3c3c2b3b7aca4e2c079"} Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.607404 4945 scope.go:117] "RemoveContainer" containerID="ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.607581 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ns5j" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.632843 4945 scope.go:117] "RemoveContainer" containerID="e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.677051 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2ns5j"] Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.678151 4945 scope.go:117] "RemoveContainer" containerID="7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.697263 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2ns5j"] Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.704232 4945 scope.go:117] "RemoveContainer" containerID="ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81" Oct 08 16:59:14 crc kubenswrapper[4945]: E1008 16:59:14.704673 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81\": container with ID starting with ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81 not found: ID does not exist" containerID="ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.704720 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81"} err="failed to get container status \"ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81\": rpc error: code = NotFound desc = could not find container \"ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81\": container with ID starting with ad40f2b8fc96045e89e45107599faf2872d454b65656ca940fd293ee3111ca81 not found: ID does not exist" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.704746 4945 scope.go:117] "RemoveContainer" containerID="e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e" Oct 08 16:59:14 crc kubenswrapper[4945]: E1008 16:59:14.707253 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e\": container with ID starting with e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e not found: ID does not exist" containerID="e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.707296 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e"} err="failed to get container status \"e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e\": rpc error: code = NotFound desc = could not find container \"e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e\": container with ID starting with e324ce33523b7a753699995031e6abbc288b805322d88a06e86869c0c05f211e not found: ID does not exist" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.707322 4945 scope.go:117] "RemoveContainer" containerID="7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9" Oct 08 16:59:14 crc kubenswrapper[4945]: E1008 16:59:14.715765 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9\": container with ID starting with 7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9 not found: ID does not exist" containerID="7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9" Oct 08 16:59:14 crc kubenswrapper[4945]: I1008 16:59:14.715799 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9"} err="failed to get container status \"7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9\": rpc error: code = NotFound desc = could not find container \"7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9\": container with ID starting with 7403c35a186a6e91355ce3c32959f2f91192402b3ac920953628fd5de33c59c9 not found: ID does not exist" Oct 08 16:59:15 crc kubenswrapper[4945]: I1008 16:59:15.680651 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ceb183ce-9008-4a1a-8b6b-26d47b2b246d/memcached/0.log" Oct 08 16:59:16 crc kubenswrapper[4945]: I1008 16:59:16.036810 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" path="/var/lib/kubelet/pods/2c7b8ca1-c98f-48e2-bbaa-0914faec510a/volumes" Oct 08 16:59:19 crc kubenswrapper[4945]: I1008 16:59:19.184480 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:59:19 crc kubenswrapper[4945]: I1008 16:59:19.185007 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:59:33 crc kubenswrapper[4945]: I1008 16:59:33.834588 4945 generic.go:334] "Generic (PLEG): container finished" podID="9fedd4a7-1a9e-4a15-a766-f54cbc78b33f" containerID="914c38fd8a3473fb366ef1b84dcabd2e44c1dd378d9e0e7020bcf767b5d981c4" exitCode=0 Oct 08 16:59:33 crc kubenswrapper[4945]: I1008 16:59:33.834791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" event={"ID":"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f","Type":"ContainerDied","Data":"914c38fd8a3473fb366ef1b84dcabd2e44c1dd378d9e0e7020bcf767b5d981c4"} Oct 08 16:59:34 crc kubenswrapper[4945]: I1008 16:59:34.990134 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.029470 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wf46x/crc-debug-gfrtq"] Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.039271 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wf46x/crc-debug-gfrtq"] Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.177319 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2v6p\" (UniqueName: \"kubernetes.io/projected/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-kube-api-access-q2v6p\") pod \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.177370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-host\") pod \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\" (UID: \"9fedd4a7-1a9e-4a15-a766-f54cbc78b33f\") " Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.177554 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-host" (OuterVolumeSpecName: "host") pod "9fedd4a7-1a9e-4a15-a766-f54cbc78b33f" (UID: "9fedd4a7-1a9e-4a15-a766-f54cbc78b33f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.178063 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-host\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.189425 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-kube-api-access-q2v6p" (OuterVolumeSpecName: "kube-api-access-q2v6p") pod "9fedd4a7-1a9e-4a15-a766-f54cbc78b33f" (UID: "9fedd4a7-1a9e-4a15-a766-f54cbc78b33f"). InnerVolumeSpecName "kube-api-access-q2v6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.279720 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2v6p\" (UniqueName: \"kubernetes.io/projected/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f-kube-api-access-q2v6p\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.859609 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05ad78ffd3e4fb8c6256e96a422d1a9f692db3a911a819e008e5938f9f146d2d" Oct 08 16:59:35 crc kubenswrapper[4945]: I1008 16:59:35.860018 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-gfrtq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.035798 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fedd4a7-1a9e-4a15-a766-f54cbc78b33f" path="/var/lib/kubelet/pods/9fedd4a7-1a9e-4a15-a766-f54cbc78b33f/volumes" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.215800 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wf46x/crc-debug-sz6lq"] Oct 08 16:59:36 crc kubenswrapper[4945]: E1008 16:59:36.216626 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="extract-utilities" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.216672 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="extract-utilities" Oct 08 16:59:36 crc kubenswrapper[4945]: E1008 16:59:36.216742 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="registry-server" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.216761 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="registry-server" Oct 08 16:59:36 crc kubenswrapper[4945]: E1008 16:59:36.216820 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fedd4a7-1a9e-4a15-a766-f54cbc78b33f" containerName="container-00" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.216842 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fedd4a7-1a9e-4a15-a766-f54cbc78b33f" containerName="container-00" Oct 08 16:59:36 crc kubenswrapper[4945]: E1008 16:59:36.216874 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="extract-content" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.216891 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="extract-content" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.217370 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7b8ca1-c98f-48e2-bbaa-0914faec510a" containerName="registry-server" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.217456 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fedd4a7-1a9e-4a15-a766-f54cbc78b33f" containerName="container-00" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.218726 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.305774 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfedf83d-d238-4c80-8daa-a3bc43d0f487-host\") pod \"crc-debug-sz6lq\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.306391 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hhvw\" (UniqueName: \"kubernetes.io/projected/cfedf83d-d238-4c80-8daa-a3bc43d0f487-kube-api-access-7hhvw\") pod \"crc-debug-sz6lq\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.408428 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hhvw\" (UniqueName: \"kubernetes.io/projected/cfedf83d-d238-4c80-8daa-a3bc43d0f487-kube-api-access-7hhvw\") pod \"crc-debug-sz6lq\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.408538 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfedf83d-d238-4c80-8daa-a3bc43d0f487-host\") pod \"crc-debug-sz6lq\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.408716 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfedf83d-d238-4c80-8daa-a3bc43d0f487-host\") pod \"crc-debug-sz6lq\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.431297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hhvw\" (UniqueName: \"kubernetes.io/projected/cfedf83d-d238-4c80-8daa-a3bc43d0f487-kube-api-access-7hhvw\") pod \"crc-debug-sz6lq\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.547280 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:36 crc kubenswrapper[4945]: I1008 16:59:36.870386 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-sz6lq" event={"ID":"cfedf83d-d238-4c80-8daa-a3bc43d0f487","Type":"ContainerStarted","Data":"905a21e7cc9ab560bae7e21bfb9d076cf846791f6bfd358969f1f8392fd0afef"} Oct 08 16:59:37 crc kubenswrapper[4945]: I1008 16:59:37.880872 4945 generic.go:334] "Generic (PLEG): container finished" podID="cfedf83d-d238-4c80-8daa-a3bc43d0f487" containerID="136894bc58be6476aafa6dc7cdf5be40e06fc3a69398944614f52fb759e0c2f2" exitCode=0 Oct 08 16:59:37 crc kubenswrapper[4945]: I1008 16:59:37.880963 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-sz6lq" event={"ID":"cfedf83d-d238-4c80-8daa-a3bc43d0f487","Type":"ContainerDied","Data":"136894bc58be6476aafa6dc7cdf5be40e06fc3a69398944614f52fb759e0c2f2"} Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.014841 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.156971 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfedf83d-d238-4c80-8daa-a3bc43d0f487-host\") pod \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.157019 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hhvw\" (UniqueName: \"kubernetes.io/projected/cfedf83d-d238-4c80-8daa-a3bc43d0f487-kube-api-access-7hhvw\") pod \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\" (UID: \"cfedf83d-d238-4c80-8daa-a3bc43d0f487\") " Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.158020 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfedf83d-d238-4c80-8daa-a3bc43d0f487-host" (OuterVolumeSpecName: "host") pod "cfedf83d-d238-4c80-8daa-a3bc43d0f487" (UID: "cfedf83d-d238-4c80-8daa-a3bc43d0f487"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.164570 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfedf83d-d238-4c80-8daa-a3bc43d0f487-kube-api-access-7hhvw" (OuterVolumeSpecName: "kube-api-access-7hhvw") pod "cfedf83d-d238-4c80-8daa-a3bc43d0f487" (UID: "cfedf83d-d238-4c80-8daa-a3bc43d0f487"). InnerVolumeSpecName "kube-api-access-7hhvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.258905 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfedf83d-d238-4c80-8daa-a3bc43d0f487-host\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.258935 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hhvw\" (UniqueName: \"kubernetes.io/projected/cfedf83d-d238-4c80-8daa-a3bc43d0f487-kube-api-access-7hhvw\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.901509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-sz6lq" event={"ID":"cfedf83d-d238-4c80-8daa-a3bc43d0f487","Type":"ContainerDied","Data":"905a21e7cc9ab560bae7e21bfb9d076cf846791f6bfd358969f1f8392fd0afef"} Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.901840 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="905a21e7cc9ab560bae7e21bfb9d076cf846791f6bfd358969f1f8392fd0afef" Oct 08 16:59:39 crc kubenswrapper[4945]: I1008 16:59:39.901604 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-sz6lq" Oct 08 16:59:47 crc kubenswrapper[4945]: I1008 16:59:47.850792 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wf46x/crc-debug-sz6lq"] Oct 08 16:59:47 crc kubenswrapper[4945]: I1008 16:59:47.860497 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wf46x/crc-debug-sz6lq"] Oct 08 16:59:48 crc kubenswrapper[4945]: I1008 16:59:48.039204 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfedf83d-d238-4c80-8daa-a3bc43d0f487" path="/var/lib/kubelet/pods/cfedf83d-d238-4c80-8daa-a3bc43d0f487/volumes" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.049789 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wf46x/crc-debug-l5q28"] Oct 08 16:59:49 crc kubenswrapper[4945]: E1008 16:59:49.050450 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfedf83d-d238-4c80-8daa-a3bc43d0f487" containerName="container-00" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.050462 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfedf83d-d238-4c80-8daa-a3bc43d0f487" containerName="container-00" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.050667 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfedf83d-d238-4c80-8daa-a3bc43d0f487" containerName="container-00" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.051372 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.184195 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.184633 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.240707 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-host\") pod \"crc-debug-l5q28\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.240855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhmt9\" (UniqueName: \"kubernetes.io/projected/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-kube-api-access-fhmt9\") pod \"crc-debug-l5q28\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.343749 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-host\") pod \"crc-debug-l5q28\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.343615 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-host\") pod \"crc-debug-l5q28\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.344093 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhmt9\" (UniqueName: \"kubernetes.io/projected/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-kube-api-access-fhmt9\") pod \"crc-debug-l5q28\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.371424 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhmt9\" (UniqueName: \"kubernetes.io/projected/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-kube-api-access-fhmt9\") pod \"crc-debug-l5q28\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.669377 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.995759 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-l5q28" event={"ID":"ab6b1e7d-4702-4553-b31c-47c68bef0ea3","Type":"ContainerStarted","Data":"6506e3415b7a962450f0dda9f71738ecd4f674b699d670f17e023915b3815b82"} Oct 08 16:59:49 crc kubenswrapper[4945]: I1008 16:59:49.996194 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-l5q28" event={"ID":"ab6b1e7d-4702-4553-b31c-47c68bef0ea3","Type":"ContainerStarted","Data":"c30c9ae454766f1ddb6ee6125b1487e6da9c38d6a8b095f4f741aa3773b1048c"} Oct 08 16:59:50 crc kubenswrapper[4945]: I1008 16:59:50.022322 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wf46x/crc-debug-l5q28" podStartSLOduration=1.02230572 podStartE2EDuration="1.02230572s" podCreationTimestamp="2025-10-08 16:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 16:59:50.012609517 +0000 UTC m=+6399.366524428" watchObservedRunningTime="2025-10-08 16:59:50.02230572 +0000 UTC m=+6399.376220621" Oct 08 16:59:51 crc kubenswrapper[4945]: I1008 16:59:51.005633 4945 generic.go:334] "Generic (PLEG): container finished" podID="ab6b1e7d-4702-4553-b31c-47c68bef0ea3" containerID="6506e3415b7a962450f0dda9f71738ecd4f674b699d670f17e023915b3815b82" exitCode=0 Oct 08 16:59:51 crc kubenswrapper[4945]: I1008 16:59:51.005710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/crc-debug-l5q28" event={"ID":"ab6b1e7d-4702-4553-b31c-47c68bef0ea3","Type":"ContainerDied","Data":"6506e3415b7a962450f0dda9f71738ecd4f674b699d670f17e023915b3815b82"} Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.140928 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.176738 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wf46x/crc-debug-l5q28"] Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.186493 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wf46x/crc-debug-l5q28"] Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.315543 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhmt9\" (UniqueName: \"kubernetes.io/projected/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-kube-api-access-fhmt9\") pod \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.315598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-host\") pod \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\" (UID: \"ab6b1e7d-4702-4553-b31c-47c68bef0ea3\") " Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.315695 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-host" (OuterVolumeSpecName: "host") pod "ab6b1e7d-4702-4553-b31c-47c68bef0ea3" (UID: "ab6b1e7d-4702-4553-b31c-47c68bef0ea3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.316440 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-host\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.329585 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-kube-api-access-fhmt9" (OuterVolumeSpecName: "kube-api-access-fhmt9") pod "ab6b1e7d-4702-4553-b31c-47c68bef0ea3" (UID: "ab6b1e7d-4702-4553-b31c-47c68bef0ea3"). InnerVolumeSpecName "kube-api-access-fhmt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 16:59:52 crc kubenswrapper[4945]: I1008 16:59:52.418781 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhmt9\" (UniqueName: \"kubernetes.io/projected/ab6b1e7d-4702-4553-b31c-47c68bef0ea3-kube-api-access-fhmt9\") on node \"crc\" DevicePath \"\"" Oct 08 16:59:53 crc kubenswrapper[4945]: I1008 16:59:53.026830 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c30c9ae454766f1ddb6ee6125b1487e6da9c38d6a8b095f4f741aa3773b1048c" Oct 08 16:59:53 crc kubenswrapper[4945]: I1008 16:59:53.026883 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/crc-debug-l5q28" Oct 08 16:59:53 crc kubenswrapper[4945]: I1008 16:59:53.950600 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/util/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.036060 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab6b1e7d-4702-4553-b31c-47c68bef0ea3" path="/var/lib/kubelet/pods/ab6b1e7d-4702-4553-b31c-47c68bef0ea3/volumes" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.152622 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/pull/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.173574 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/util/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.181003 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/pull/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.358015 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/util/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.379754 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/extract/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.397101 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a35df681a0d9f8331d5b70b00d0344256b38289275b54ac1f9a9331148fz9bl_00cf9b93-fcae-4759-936e-58397840df0a/pull/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.535585 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-7b8pm_2747410b-55fa-4a7f-9146-5662ee48cce7/kube-rbac-proxy/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.594967 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-7b8pm_2747410b-55fa-4a7f-9146-5662ee48cce7/manager/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.627989 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-h27l8_441d050d-f359-46d8-8c29-0b1506c1f240/kube-rbac-proxy/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.779678 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-h27l8_441d050d-f359-46d8-8c29-0b1506c1f240/manager/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.829652 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-4fjrf_c3ae7aef-2f32-4d2d-b8be-e29dd91ede71/kube-rbac-proxy/0.log" Oct 08 16:59:54 crc kubenswrapper[4945]: I1008 16:59:54.836200 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-4fjrf_c3ae7aef-2f32-4d2d-b8be-e29dd91ede71/manager/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.028895 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-kx5k7_5c314cb3-27af-4a08-b96b-91f9a70f7524/kube-rbac-proxy/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.103685 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-kx5k7_5c314cb3-27af-4a08-b96b-91f9a70f7524/manager/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.228413 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-62hpj_84979368-496d-43c1-b792-20c167a2e5d5/manager/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.248376 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-62hpj_84979368-496d-43c1-b792-20c167a2e5d5/kube-rbac-proxy/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.329341 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-j9vft_b32201a1-e841-4c4c-8988-0b866243b255/kube-rbac-proxy/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.430779 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-j9vft_b32201a1-e841-4c4c-8988-0b866243b255/manager/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.496815 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-r9mp8_f11222c0-9d0f-4d3e-945f-b991135e4235/kube-rbac-proxy/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.671602 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-hk5rr_25cc3669-bdfd-44eb-83d0-85ca303250ea/kube-rbac-proxy/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.731592 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-r9mp8_f11222c0-9d0f-4d3e-945f-b991135e4235/manager/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.753022 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-hk5rr_25cc3669-bdfd-44eb-83d0-85ca303250ea/manager/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.870117 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-lr6hn_344902f1-cf0c-4390-8f2c-2fe41cf9ccd9/kube-rbac-proxy/0.log" Oct 08 16:59:55 crc kubenswrapper[4945]: I1008 16:59:55.989933 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-lr6hn_344902f1-cf0c-4390-8f2c-2fe41cf9ccd9/manager/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.038302 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ht7cm_cbe3a47b-033a-410d-b2be-bd38e4b9c0ab/kube-rbac-proxy/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.086849 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ht7cm_cbe3a47b-033a-410d-b2be-bd38e4b9c0ab/manager/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.193356 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-p7bfn_b2ca2442-7e23-40a8-9ca1-1ef11251d34d/kube-rbac-proxy/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.267588 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-p7bfn_b2ca2442-7e23-40a8-9ca1-1ef11251d34d/manager/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.363904 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-t56f2_1112a081-2d10-48e2-b345-8ef841c5dc50/kube-rbac-proxy/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.460296 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-t56f2_1112a081-2d10-48e2-b345-8ef841c5dc50/manager/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.504954 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-lwppj_3501f2c3-1b0d-4f7a-9fe2-2516f0088993/kube-rbac-proxy/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.603838 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-lwppj_3501f2c3-1b0d-4f7a-9fe2-2516f0088993/manager/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.687289 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-tb4rv_9555aa6e-7b8f-4354-96e8-e639719e1b56/kube-rbac-proxy/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.715184 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-tb4rv_9555aa6e-7b8f-4354-96e8-e639719e1b56/manager/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.849960 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt_a8d00f8c-42c6-4d99-bb76-bdd491597e3a/kube-rbac-proxy/0.log" Oct 08 16:59:56 crc kubenswrapper[4945]: I1008 16:59:56.870810 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d6flkt_a8d00f8c-42c6-4d99-bb76-bdd491597e3a/manager/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.039996 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56688bc7d6-h2wfx_c42bd8a4-d1ed-446c-9295-03cfe33588de/kube-rbac-proxy/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.178552 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75967d884c-qslk5_ef574f9e-15f9-42fe-8593-5575c31fe0d9/kube-rbac-proxy/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.378917 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75967d884c-qslk5_ef574f9e-15f9-42fe-8593-5575c31fe0d9/operator/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.436777 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-65hg6_9844bfac-ebec-4692-abda-eacfe73d5b3a/registry-server/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.601794 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-5t8fx_400ef757-79e4-4dd6-8c0a-2440835ac048/kube-rbac-proxy/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.687745 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-5t8fx_400ef757-79e4-4dd6-8c0a-2440835ac048/manager/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.785266 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-c8h57_efd39ce1-ac6e-449d-b32f-fc451ffde97e/kube-rbac-proxy/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.834310 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-c8h57_efd39ce1-ac6e-449d-b32f-fc451ffde97e/manager/0.log" Oct 08 16:59:57 crc kubenswrapper[4945]: I1008 16:59:57.988376 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-rrjcx_1d068bb6-a523-4903-95b7-7c62862c2bb3/operator/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.092785 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qxbqr_43c420bf-cf21-4caa-90ce-0a9009d3bc45/kube-rbac-proxy/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.248496 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qxbqr_43c420bf-cf21-4caa-90ce-0a9009d3bc45/manager/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.275898 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-rn55t_f8b50124-bf7a-497b-bdc6-42bbbe564a7d/kube-rbac-proxy/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.511241 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-dwfj9_28b9d7cf-ad46-4c2a-a942-f7b20c054216/kube-rbac-proxy/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.521305 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-dwfj9_28b9d7cf-ad46-4c2a-a942-f7b20c054216/manager/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.521659 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-56688bc7d6-h2wfx_c42bd8a4-d1ed-446c-9295-03cfe33588de/manager/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.658877 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-rn55t_f8b50124-bf7a-497b-bdc6-42bbbe564a7d/manager/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.743879 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-df6684d58-22z48_e0437b00-8fcc-412a-90fa-32ea56c71421/kube-rbac-proxy/0.log" Oct 08 16:59:58 crc kubenswrapper[4945]: I1008 16:59:58.822652 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-df6684d58-22z48_e0437b00-8fcc-412a-90fa-32ea56c71421/manager/0.log" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.164366 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb"] Oct 08 17:00:00 crc kubenswrapper[4945]: E1008 17:00:00.165258 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab6b1e7d-4702-4553-b31c-47c68bef0ea3" containerName="container-00" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.165275 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab6b1e7d-4702-4553-b31c-47c68bef0ea3" containerName="container-00" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.165509 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab6b1e7d-4702-4553-b31c-47c68bef0ea3" containerName="container-00" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.166395 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.169306 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.169569 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.172955 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb"] Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.199818 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39e01145-b997-4c7e-874c-c092f68e15f3-secret-volume\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.199923 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39e01145-b997-4c7e-874c-c092f68e15f3-config-volume\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.199956 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snm6r\" (UniqueName: \"kubernetes.io/projected/39e01145-b997-4c7e-874c-c092f68e15f3-kube-api-access-snm6r\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.302257 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39e01145-b997-4c7e-874c-c092f68e15f3-config-volume\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.302317 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snm6r\" (UniqueName: \"kubernetes.io/projected/39e01145-b997-4c7e-874c-c092f68e15f3-kube-api-access-snm6r\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.302449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39e01145-b997-4c7e-874c-c092f68e15f3-secret-volume\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.303140 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39e01145-b997-4c7e-874c-c092f68e15f3-config-volume\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.310631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39e01145-b997-4c7e-874c-c092f68e15f3-secret-volume\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.324554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snm6r\" (UniqueName: \"kubernetes.io/projected/39e01145-b997-4c7e-874c-c092f68e15f3-kube-api-access-snm6r\") pod \"collect-profiles-29332380-4f8tb\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.494603 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:00 crc kubenswrapper[4945]: I1008 17:00:00.979573 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb"] Oct 08 17:00:01 crc kubenswrapper[4945]: I1008 17:00:01.098806 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" event={"ID":"39e01145-b997-4c7e-874c-c092f68e15f3","Type":"ContainerStarted","Data":"70616d93cda625e328f9ce5eee012f5e50b4ca83a2e88431553ccd4a21c56b3d"} Oct 08 17:00:02 crc kubenswrapper[4945]: I1008 17:00:02.111485 4945 generic.go:334] "Generic (PLEG): container finished" podID="39e01145-b997-4c7e-874c-c092f68e15f3" containerID="22eabb1e45cc4045f9ecc73160a4e043f7dcd1c54b8c33e32b4b382891d88010" exitCode=0 Oct 08 17:00:02 crc kubenswrapper[4945]: I1008 17:00:02.111657 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" event={"ID":"39e01145-b997-4c7e-874c-c092f68e15f3","Type":"ContainerDied","Data":"22eabb1e45cc4045f9ecc73160a4e043f7dcd1c54b8c33e32b4b382891d88010"} Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.466586 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.574221 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39e01145-b997-4c7e-874c-c092f68e15f3-secret-volume\") pod \"39e01145-b997-4c7e-874c-c092f68e15f3\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.574355 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39e01145-b997-4c7e-874c-c092f68e15f3-config-volume\") pod \"39e01145-b997-4c7e-874c-c092f68e15f3\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.574589 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snm6r\" (UniqueName: \"kubernetes.io/projected/39e01145-b997-4c7e-874c-c092f68e15f3-kube-api-access-snm6r\") pod \"39e01145-b997-4c7e-874c-c092f68e15f3\" (UID: \"39e01145-b997-4c7e-874c-c092f68e15f3\") " Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.574867 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e01145-b997-4c7e-874c-c092f68e15f3-config-volume" (OuterVolumeSpecName: "config-volume") pod "39e01145-b997-4c7e-874c-c092f68e15f3" (UID: "39e01145-b997-4c7e-874c-c092f68e15f3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.575229 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39e01145-b997-4c7e-874c-c092f68e15f3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.581505 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e01145-b997-4c7e-874c-c092f68e15f3-kube-api-access-snm6r" (OuterVolumeSpecName: "kube-api-access-snm6r") pod "39e01145-b997-4c7e-874c-c092f68e15f3" (UID: "39e01145-b997-4c7e-874c-c092f68e15f3"). InnerVolumeSpecName "kube-api-access-snm6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.588300 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39e01145-b997-4c7e-874c-c092f68e15f3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "39e01145-b997-4c7e-874c-c092f68e15f3" (UID: "39e01145-b997-4c7e-874c-c092f68e15f3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.677103 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snm6r\" (UniqueName: \"kubernetes.io/projected/39e01145-b997-4c7e-874c-c092f68e15f3-kube-api-access-snm6r\") on node \"crc\" DevicePath \"\"" Oct 08 17:00:03 crc kubenswrapper[4945]: I1008 17:00:03.677205 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39e01145-b997-4c7e-874c-c092f68e15f3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 17:00:04 crc kubenswrapper[4945]: I1008 17:00:04.133759 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" event={"ID":"39e01145-b997-4c7e-874c-c092f68e15f3","Type":"ContainerDied","Data":"70616d93cda625e328f9ce5eee012f5e50b4ca83a2e88431553ccd4a21c56b3d"} Oct 08 17:00:04 crc kubenswrapper[4945]: I1008 17:00:04.133794 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70616d93cda625e328f9ce5eee012f5e50b4ca83a2e88431553ccd4a21c56b3d" Oct 08 17:00:04 crc kubenswrapper[4945]: I1008 17:00:04.133797 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332380-4f8tb" Oct 08 17:00:04 crc kubenswrapper[4945]: I1008 17:00:04.552362 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm"] Oct 08 17:00:04 crc kubenswrapper[4945]: I1008 17:00:04.560370 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332335-whpgm"] Oct 08 17:00:06 crc kubenswrapper[4945]: I1008 17:00:06.036934 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f09c79-5895-41ac-830e-eb717471c62b" path="/var/lib/kubelet/pods/e7f09c79-5895-41ac-830e-eb717471c62b/volumes" Oct 08 17:00:13 crc kubenswrapper[4945]: I1008 17:00:13.562006 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mn52j_f973b44d-6296-408b-9d14-31d56fdd1021/control-plane-machine-set-operator/0.log" Oct 08 17:00:13 crc kubenswrapper[4945]: I1008 17:00:13.733932 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j72zk_ec002292-f388-497c-b114-fb7818730f4d/kube-rbac-proxy/0.log" Oct 08 17:00:13 crc kubenswrapper[4945]: I1008 17:00:13.751573 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j72zk_ec002292-f388-497c-b114-fb7818730f4d/machine-api-operator/0.log" Oct 08 17:00:19 crc kubenswrapper[4945]: I1008 17:00:19.184445 4945 patch_prober.go:28] interesting pod/machine-config-daemon-jp64g container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 17:00:19 crc kubenswrapper[4945]: I1008 17:00:19.185009 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 17:00:19 crc kubenswrapper[4945]: I1008 17:00:19.185082 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" Oct 08 17:00:19 crc kubenswrapper[4945]: I1008 17:00:19.185901 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6"} pod="openshift-machine-config-operator/machine-config-daemon-jp64g" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 17:00:19 crc kubenswrapper[4945]: I1008 17:00:19.186043 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerName="machine-config-daemon" containerID="cri-o://30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" gracePeriod=600 Oct 08 17:00:19 crc kubenswrapper[4945]: E1008 17:00:19.309104 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:00:19 crc kubenswrapper[4945]: E1008 17:00:19.436674 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode39fc950_7f86_4d6b_b4b0_25daafb66250.slice/crio-conmon-30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode39fc950_7f86_4d6b_b4b0_25daafb66250.slice/crio-30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6.scope\": RecentStats: unable to find data in memory cache]" Oct 08 17:00:20 crc kubenswrapper[4945]: I1008 17:00:20.271742 4945 generic.go:334] "Generic (PLEG): container finished" podID="e39fc950-7f86-4d6b-b4b0-25daafb66250" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" exitCode=0 Oct 08 17:00:20 crc kubenswrapper[4945]: I1008 17:00:20.271791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerDied","Data":"30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6"} Oct 08 17:00:20 crc kubenswrapper[4945]: I1008 17:00:20.272082 4945 scope.go:117] "RemoveContainer" containerID="237f1f106bec3e491f7495488dc33c96187402c9198bafca4b532ff6e25ad8c3" Oct 08 17:00:20 crc kubenswrapper[4945]: I1008 17:00:20.273054 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:00:20 crc kubenswrapper[4945]: E1008 17:00:20.273523 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:00:25 crc kubenswrapper[4945]: I1008 17:00:25.285737 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xhl5p_8a063d76-5a7a-49f7-b19c-fac4fd303a90/cert-manager-controller/0.log" Oct 08 17:00:25 crc kubenswrapper[4945]: I1008 17:00:25.479523 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-2pms8_a020faff-9bc7-4f4c-83a4-9e21f352e5f2/cert-manager-cainjector/0.log" Oct 08 17:00:25 crc kubenswrapper[4945]: I1008 17:00:25.579340 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-q9846_ec83ad94-ee5b-4984-8381-bda1289c4ef8/cert-manager-webhook/0.log" Oct 08 17:00:31 crc kubenswrapper[4945]: I1008 17:00:31.024196 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:00:31 crc kubenswrapper[4945]: E1008 17:00:31.024910 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:00:38 crc kubenswrapper[4945]: I1008 17:00:38.252663 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-7zpln_d4d9183c-cff3-4bed-82e0-41a654969322/nmstate-console-plugin/0.log" Oct 08 17:00:38 crc kubenswrapper[4945]: I1008 17:00:38.319975 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6q4xl_835b6fb9-28f9-4c64-8049-05dd404be8cc/nmstate-handler/0.log" Oct 08 17:00:38 crc kubenswrapper[4945]: I1008 17:00:38.368751 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-kfqmx_0afd8411-7a96-44b9-b1af-82a05419fa24/kube-rbac-proxy/0.log" Oct 08 17:00:38 crc kubenswrapper[4945]: I1008 17:00:38.457724 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-kfqmx_0afd8411-7a96-44b9-b1af-82a05419fa24/nmstate-metrics/0.log" Oct 08 17:00:38 crc kubenswrapper[4945]: I1008 17:00:38.573237 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-z75g2_9d7e5f1a-5f1b-4291-8dee-de24611c5ff7/nmstate-operator/0.log" Oct 08 17:00:38 crc kubenswrapper[4945]: I1008 17:00:38.653917 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-59hrm_c84bb408-d873-4b03-a07b-dfb913fdc64b/nmstate-webhook/0.log" Oct 08 17:00:43 crc kubenswrapper[4945]: I1008 17:00:43.024483 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:00:43 crc kubenswrapper[4945]: E1008 17:00:43.025329 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:00:47 crc kubenswrapper[4945]: I1008 17:00:47.650994 4945 scope.go:117] "RemoveContainer" containerID="832a4c74370750c767b7575d8dbf1ce38f0bb75cc2317ae337815f30e9582d3b" Oct 08 17:00:52 crc kubenswrapper[4945]: I1008 17:00:52.860163 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-nnbmf_919b5756-8417-47d1-acd7-fbfd2c01d246/controller/0.log" Oct 08 17:00:52 crc kubenswrapper[4945]: I1008 17:00:52.868810 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-nnbmf_919b5756-8417-47d1-acd7-fbfd2c01d246/kube-rbac-proxy/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.052677 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.217291 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.223760 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.227546 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.301643 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.505940 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.521718 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.547081 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.568403 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.727222 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-frr-files/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.750878 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-reloader/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.751053 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/cp-metrics/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.788817 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/controller/0.log" Oct 08 17:00:53 crc kubenswrapper[4945]: I1008 17:00:53.918203 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/frr-metrics/0.log" Oct 08 17:00:54 crc kubenswrapper[4945]: I1008 17:00:54.017962 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/kube-rbac-proxy/0.log" Oct 08 17:00:54 crc kubenswrapper[4945]: I1008 17:00:54.027691 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/kube-rbac-proxy-frr/0.log" Oct 08 17:00:54 crc kubenswrapper[4945]: I1008 17:00:54.159281 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/reloader/0.log" Oct 08 17:00:54 crc kubenswrapper[4945]: I1008 17:00:54.269576 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-h5s5z_1ae8dfbc-86f8-4b14-a8a3-42d6e7f12b3c/frr-k8s-webhook-server/0.log" Oct 08 17:00:54 crc kubenswrapper[4945]: I1008 17:00:54.512701 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-575bbc8666-4t674_13c15125-a872-459e-ac9a-d37356f47418/manager/0.log" Oct 08 17:00:54 crc kubenswrapper[4945]: I1008 17:00:54.605831 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d999db554-m5t6x_132bc709-76d1-41fc-ae20-04141e80a6ac/webhook-server/0.log" Oct 08 17:00:54 crc kubenswrapper[4945]: I1008 17:00:54.878052 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dftb8_283c80b9-bfe4-4c26-852d-9ba6d14009d4/kube-rbac-proxy/0.log" Oct 08 17:00:55 crc kubenswrapper[4945]: I1008 17:00:55.415137 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dftb8_283c80b9-bfe4-4c26-852d-9ba6d14009d4/speaker/0.log" Oct 08 17:00:55 crc kubenswrapper[4945]: I1008 17:00:55.921018 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cc5ps_c0c94cdf-157a-48b2-b8ff-389b2ec3e04d/frr/0.log" Oct 08 17:00:57 crc kubenswrapper[4945]: I1008 17:00:57.024381 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:00:57 crc kubenswrapper[4945]: E1008 17:00:57.024659 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.144774 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29332381-tbsbf"] Oct 08 17:01:00 crc kubenswrapper[4945]: E1008 17:01:00.145724 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e01145-b997-4c7e-874c-c092f68e15f3" containerName="collect-profiles" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.145742 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e01145-b997-4c7e-874c-c092f68e15f3" containerName="collect-profiles" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.146005 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="39e01145-b997-4c7e-874c-c092f68e15f3" containerName="collect-profiles" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.146927 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.156483 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332381-tbsbf"] Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.263762 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-config-data\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.263958 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-combined-ca-bundle\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.264005 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfbgv\" (UniqueName: \"kubernetes.io/projected/06e66921-2c13-49e5-ac1e-09122695fdc9-kube-api-access-pfbgv\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.264240 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-fernet-keys\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.365855 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-config-data\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.365979 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-combined-ca-bundle\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.366008 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfbgv\" (UniqueName: \"kubernetes.io/projected/06e66921-2c13-49e5-ac1e-09122695fdc9-kube-api-access-pfbgv\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.366047 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-fernet-keys\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.372850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-combined-ca-bundle\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.386263 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-fernet-keys\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.386351 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-config-data\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.389812 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfbgv\" (UniqueName: \"kubernetes.io/projected/06e66921-2c13-49e5-ac1e-09122695fdc9-kube-api-access-pfbgv\") pod \"keystone-cron-29332381-tbsbf\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.468420 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:00 crc kubenswrapper[4945]: I1008 17:01:00.941295 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332381-tbsbf"] Oct 08 17:01:01 crc kubenswrapper[4945]: I1008 17:01:01.670915 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332381-tbsbf" event={"ID":"06e66921-2c13-49e5-ac1e-09122695fdc9","Type":"ContainerStarted","Data":"55cb3f11335ac45e152c285ac4b7feb8ba53bf6668f45992e17adc2c06b5023a"} Oct 08 17:01:01 crc kubenswrapper[4945]: I1008 17:01:01.672380 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332381-tbsbf" event={"ID":"06e66921-2c13-49e5-ac1e-09122695fdc9","Type":"ContainerStarted","Data":"56cec1ded33f0bb7834aaf0574f08a69d411eb35c14ecf58092b9e6358524b8a"} Oct 08 17:01:01 crc kubenswrapper[4945]: I1008 17:01:01.698868 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29332381-tbsbf" podStartSLOduration=1.698845054 podStartE2EDuration="1.698845054s" podCreationTimestamp="2025-10-08 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 17:01:01.693774497 +0000 UTC m=+6471.047689398" watchObservedRunningTime="2025-10-08 17:01:01.698845054 +0000 UTC m=+6471.052759955" Oct 08 17:01:05 crc kubenswrapper[4945]: I1008 17:01:05.719244 4945 generic.go:334] "Generic (PLEG): container finished" podID="06e66921-2c13-49e5-ac1e-09122695fdc9" containerID="55cb3f11335ac45e152c285ac4b7feb8ba53bf6668f45992e17adc2c06b5023a" exitCode=0 Oct 08 17:01:05 crc kubenswrapper[4945]: I1008 17:01:05.719314 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332381-tbsbf" event={"ID":"06e66921-2c13-49e5-ac1e-09122695fdc9","Type":"ContainerDied","Data":"55cb3f11335ac45e152c285ac4b7feb8ba53bf6668f45992e17adc2c06b5023a"} Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.123828 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.214162 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfbgv\" (UniqueName: \"kubernetes.io/projected/06e66921-2c13-49e5-ac1e-09122695fdc9-kube-api-access-pfbgv\") pod \"06e66921-2c13-49e5-ac1e-09122695fdc9\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.214227 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-fernet-keys\") pod \"06e66921-2c13-49e5-ac1e-09122695fdc9\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.214348 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-combined-ca-bundle\") pod \"06e66921-2c13-49e5-ac1e-09122695fdc9\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.214485 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-config-data\") pod \"06e66921-2c13-49e5-ac1e-09122695fdc9\" (UID: \"06e66921-2c13-49e5-ac1e-09122695fdc9\") " Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.221314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "06e66921-2c13-49e5-ac1e-09122695fdc9" (UID: "06e66921-2c13-49e5-ac1e-09122695fdc9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.236036 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e66921-2c13-49e5-ac1e-09122695fdc9-kube-api-access-pfbgv" (OuterVolumeSpecName: "kube-api-access-pfbgv") pod "06e66921-2c13-49e5-ac1e-09122695fdc9" (UID: "06e66921-2c13-49e5-ac1e-09122695fdc9"). InnerVolumeSpecName "kube-api-access-pfbgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.257889 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06e66921-2c13-49e5-ac1e-09122695fdc9" (UID: "06e66921-2c13-49e5-ac1e-09122695fdc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.275282 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-config-data" (OuterVolumeSpecName: "config-data") pod "06e66921-2c13-49e5-ac1e-09122695fdc9" (UID: "06e66921-2c13-49e5-ac1e-09122695fdc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.317304 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfbgv\" (UniqueName: \"kubernetes.io/projected/06e66921-2c13-49e5-ac1e-09122695fdc9-kube-api-access-pfbgv\") on node \"crc\" DevicePath \"\"" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.317347 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.317361 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.317372 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e66921-2c13-49e5-ac1e-09122695fdc9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.742887 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332381-tbsbf" event={"ID":"06e66921-2c13-49e5-ac1e-09122695fdc9","Type":"ContainerDied","Data":"56cec1ded33f0bb7834aaf0574f08a69d411eb35c14ecf58092b9e6358524b8a"} Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.742932 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56cec1ded33f0bb7834aaf0574f08a69d411eb35c14ecf58092b9e6358524b8a" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.742950 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332381-tbsbf" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.800790 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/util/0.log" Oct 08 17:01:07 crc kubenswrapper[4945]: I1008 17:01:07.988298 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/util/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.006085 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/pull/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.034168 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/pull/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.167406 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/pull/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.205362 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/util/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.219640 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2876sl_b6c23df0-35e1-4905-bf5a-84b240ea3b0d/extract/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.322219 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/util/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.532681 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/pull/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.552198 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/pull/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.578172 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/util/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.749196 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/pull/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.768626 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/extract/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.797180 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwsjks_81f62690-bcd9-4d49-ac08-c754f28e143b/util/0.log" Oct 08 17:01:08 crc kubenswrapper[4945]: I1008 17:01:08.921331 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-utilities/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.093322 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-utilities/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.093484 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-content/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.130987 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-content/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.290732 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-utilities/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.383769 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/extract-content/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.503749 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-utilities/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.769017 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-content/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.796706 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-content/0.log" Oct 08 17:01:09 crc kubenswrapper[4945]: I1008 17:01:09.816388 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-utilities/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.129834 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-utilities/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.158881 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/extract-content/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.169917 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cxqrl_a7bde639-beb8-4042-9adc-ad1dcde01cfc/registry-server/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.417934 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/util/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.593817 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/util/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.692424 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/pull/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.702483 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/pull/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.837235 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/util/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.913709 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/pull/0.log" Oct 08 17:01:10 crc kubenswrapper[4945]: I1008 17:01:10.940524 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjm6mb_9b9fcc55-1a43-4136-aebc-50597faeb89d/extract/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.126560 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rzsp2_9c54859f-2831-49c8-88db-4b0c4451f3b0/registry-server/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.138624 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-vnm62_e99374e4-2a46-4044-881c-9bdb32f4a215/marketplace-operator/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.297221 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-utilities/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.501544 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-utilities/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.501602 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-content/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.519960 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-content/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.653260 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-utilities/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.682741 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/extract-content/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.743687 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-utilities/0.log" Oct 08 17:01:11 crc kubenswrapper[4945]: I1008 17:01:11.907681 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-64lj5_6ff2444c-a4f4-4779-a11c-968c0ed65815/registry-server/0.log" Oct 08 17:01:12 crc kubenswrapper[4945]: I1008 17:01:12.001875 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-content/0.log" Oct 08 17:01:12 crc kubenswrapper[4945]: I1008 17:01:12.002011 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-utilities/0.log" Oct 08 17:01:12 crc kubenswrapper[4945]: I1008 17:01:12.016778 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-content/0.log" Oct 08 17:01:12 crc kubenswrapper[4945]: I1008 17:01:12.038207 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:01:12 crc kubenswrapper[4945]: E1008 17:01:12.038527 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:01:12 crc kubenswrapper[4945]: I1008 17:01:12.212215 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-utilities/0.log" Oct 08 17:01:12 crc kubenswrapper[4945]: I1008 17:01:12.226338 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/extract-content/0.log" Oct 08 17:01:13 crc kubenswrapper[4945]: I1008 17:01:13.067475 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9xdxb_db0ecadd-b135-46e5-aa0e-92d952913a11/registry-server/0.log" Oct 08 17:01:23 crc kubenswrapper[4945]: I1008 17:01:23.644910 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-g4w8k_18169f3c-d245-4b63-8ea7-d6d3bd70cce4/prometheus-operator/0.log" Oct 08 17:01:23 crc kubenswrapper[4945]: I1008 17:01:23.814288 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6cf547974b-fpjtp_658dd3b7-2709-4785-aa43-64f041d46b7a/prometheus-operator-admission-webhook/0.log" Oct 08 17:01:23 crc kubenswrapper[4945]: I1008 17:01:23.856635 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6cf547974b-xrjkd_fed6f952-f438-4a3b-bd8c-f2832a1c6e8d/prometheus-operator-admission-webhook/0.log" Oct 08 17:01:24 crc kubenswrapper[4945]: I1008 17:01:24.028042 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-qrx7m_09a37473-34c8-43ff-a839-3f712f3d0605/operator/0.log" Oct 08 17:01:24 crc kubenswrapper[4945]: I1008 17:01:24.045186 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-sw8w6_d39d5733-b343-4a33-866b-220a8191eda9/perses-operator/0.log" Oct 08 17:01:26 crc kubenswrapper[4945]: I1008 17:01:26.024065 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:01:26 crc kubenswrapper[4945]: E1008 17:01:26.024605 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:01:39 crc kubenswrapper[4945]: I1008 17:01:39.024597 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:01:39 crc kubenswrapper[4945]: E1008 17:01:39.025560 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:01:52 crc kubenswrapper[4945]: I1008 17:01:52.031008 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:01:52 crc kubenswrapper[4945]: E1008 17:01:52.031795 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:02:04 crc kubenswrapper[4945]: I1008 17:02:04.025722 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:02:04 crc kubenswrapper[4945]: E1008 17:02:04.026496 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:02:15 crc kubenswrapper[4945]: I1008 17:02:15.024327 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:02:15 crc kubenswrapper[4945]: E1008 17:02:15.026842 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:02:30 crc kubenswrapper[4945]: I1008 17:02:30.024883 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:02:30 crc kubenswrapper[4945]: E1008 17:02:30.025740 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:02:43 crc kubenswrapper[4945]: I1008 17:02:43.024832 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:02:43 crc kubenswrapper[4945]: E1008 17:02:43.025820 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:02:56 crc kubenswrapper[4945]: I1008 17:02:56.024885 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:02:56 crc kubenswrapper[4945]: E1008 17:02:56.026164 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:03:08 crc kubenswrapper[4945]: I1008 17:03:08.023920 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:03:08 crc kubenswrapper[4945]: E1008 17:03:08.024710 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:03:23 crc kubenswrapper[4945]: I1008 17:03:23.024829 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:03:23 crc kubenswrapper[4945]: E1008 17:03:23.026143 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:03:34 crc kubenswrapper[4945]: I1008 17:03:34.024325 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:03:34 crc kubenswrapper[4945]: E1008 17:03:34.024953 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:03:47 crc kubenswrapper[4945]: I1008 17:03:47.758973 4945 scope.go:117] "RemoveContainer" containerID="914c38fd8a3473fb366ef1b84dcabd2e44c1dd378d9e0e7020bcf767b5d981c4" Oct 08 17:03:48 crc kubenswrapper[4945]: I1008 17:03:48.024180 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:03:48 crc kubenswrapper[4945]: E1008 17:03:48.024679 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:04:00 crc kubenswrapper[4945]: I1008 17:04:00.029397 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:04:00 crc kubenswrapper[4945]: E1008 17:04:00.029948 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:04:00 crc kubenswrapper[4945]: I1008 17:04:00.595252 4945 generic.go:334] "Generic (PLEG): container finished" podID="efe6cd45-99e7-42ec-8acd-c42cc681a319" containerID="c1d6d83a6ef1940a5cc25840cb95ffb6265cb5b51d050d2a72e83cc47eefbb1a" exitCode=0 Oct 08 17:04:00 crc kubenswrapper[4945]: I1008 17:04:00.595292 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wf46x/must-gather-klcdw" event={"ID":"efe6cd45-99e7-42ec-8acd-c42cc681a319","Type":"ContainerDied","Data":"c1d6d83a6ef1940a5cc25840cb95ffb6265cb5b51d050d2a72e83cc47eefbb1a"} Oct 08 17:04:00 crc kubenswrapper[4945]: I1008 17:04:00.595914 4945 scope.go:117] "RemoveContainer" containerID="c1d6d83a6ef1940a5cc25840cb95ffb6265cb5b51d050d2a72e83cc47eefbb1a" Oct 08 17:04:01 crc kubenswrapper[4945]: I1008 17:04:01.300664 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wf46x_must-gather-klcdw_efe6cd45-99e7-42ec-8acd-c42cc681a319/gather/0.log" Oct 08 17:04:14 crc kubenswrapper[4945]: I1008 17:04:14.490165 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wf46x/must-gather-klcdw"] Oct 08 17:04:14 crc kubenswrapper[4945]: I1008 17:04:14.491279 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wf46x/must-gather-klcdw" podUID="efe6cd45-99e7-42ec-8acd-c42cc681a319" containerName="copy" containerID="cri-o://dc8b0657328ec206321bc98de7304ea09705e8be01a083932d93cb11c8f29cd0" gracePeriod=2 Oct 08 17:04:14 crc kubenswrapper[4945]: I1008 17:04:14.504098 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wf46x/must-gather-klcdw"] Oct 08 17:04:14 crc kubenswrapper[4945]: I1008 17:04:14.782219 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wf46x_must-gather-klcdw_efe6cd45-99e7-42ec-8acd-c42cc681a319/copy/0.log" Oct 08 17:04:14 crc kubenswrapper[4945]: I1008 17:04:14.782607 4945 generic.go:334] "Generic (PLEG): container finished" podID="efe6cd45-99e7-42ec-8acd-c42cc681a319" containerID="dc8b0657328ec206321bc98de7304ea09705e8be01a083932d93cb11c8f29cd0" exitCode=143 Oct 08 17:04:14 crc kubenswrapper[4945]: I1008 17:04:14.974943 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wf46x_must-gather-klcdw_efe6cd45-99e7-42ec-8acd-c42cc681a319/copy/0.log" Oct 08 17:04:14 crc kubenswrapper[4945]: I1008 17:04:14.975461 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.025217 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:04:15 crc kubenswrapper[4945]: E1008 17:04:15.025559 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.125514 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efe6cd45-99e7-42ec-8acd-c42cc681a319-must-gather-output\") pod \"efe6cd45-99e7-42ec-8acd-c42cc681a319\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.125899 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vn7d\" (UniqueName: \"kubernetes.io/projected/efe6cd45-99e7-42ec-8acd-c42cc681a319-kube-api-access-2vn7d\") pod \"efe6cd45-99e7-42ec-8acd-c42cc681a319\" (UID: \"efe6cd45-99e7-42ec-8acd-c42cc681a319\") " Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.131269 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe6cd45-99e7-42ec-8acd-c42cc681a319-kube-api-access-2vn7d" (OuterVolumeSpecName: "kube-api-access-2vn7d") pod "efe6cd45-99e7-42ec-8acd-c42cc681a319" (UID: "efe6cd45-99e7-42ec-8acd-c42cc681a319"). InnerVolumeSpecName "kube-api-access-2vn7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.228507 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vn7d\" (UniqueName: \"kubernetes.io/projected/efe6cd45-99e7-42ec-8acd-c42cc681a319-kube-api-access-2vn7d\") on node \"crc\" DevicePath \"\"" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.334310 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efe6cd45-99e7-42ec-8acd-c42cc681a319-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "efe6cd45-99e7-42ec-8acd-c42cc681a319" (UID: "efe6cd45-99e7-42ec-8acd-c42cc681a319"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.432506 4945 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efe6cd45-99e7-42ec-8acd-c42cc681a319-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.797032 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wf46x_must-gather-klcdw_efe6cd45-99e7-42ec-8acd-c42cc681a319/copy/0.log" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.797636 4945 scope.go:117] "RemoveContainer" containerID="dc8b0657328ec206321bc98de7304ea09705e8be01a083932d93cb11c8f29cd0" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.797957 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wf46x/must-gather-klcdw" Oct 08 17:04:15 crc kubenswrapper[4945]: I1008 17:04:15.823756 4945 scope.go:117] "RemoveContainer" containerID="c1d6d83a6ef1940a5cc25840cb95ffb6265cb5b51d050d2a72e83cc47eefbb1a" Oct 08 17:04:15 crc kubenswrapper[4945]: E1008 17:04:15.981800 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefe6cd45_99e7_42ec_8acd_c42cc681a319.slice\": RecentStats: unable to find data in memory cache]" Oct 08 17:04:16 crc kubenswrapper[4945]: I1008 17:04:16.036684 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe6cd45-99e7-42ec-8acd-c42cc681a319" path="/var/lib/kubelet/pods/efe6cd45-99e7-42ec-8acd-c42cc681a319/volumes" Oct 08 17:04:30 crc kubenswrapper[4945]: I1008 17:04:30.024811 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:04:30 crc kubenswrapper[4945]: E1008 17:04:30.025908 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:04:42 crc kubenswrapper[4945]: I1008 17:04:42.031447 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:04:42 crc kubenswrapper[4945]: E1008 17:04:42.032920 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:04:53 crc kubenswrapper[4945]: I1008 17:04:53.024028 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:04:53 crc kubenswrapper[4945]: E1008 17:04:53.025034 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:05:08 crc kubenswrapper[4945]: I1008 17:05:08.024889 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:05:08 crc kubenswrapper[4945]: E1008 17:05:08.026309 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jp64g_openshift-machine-config-operator(e39fc950-7f86-4d6b-b4b0-25daafb66250)\"" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" podUID="e39fc950-7f86-4d6b-b4b0-25daafb66250" Oct 08 17:05:20 crc kubenswrapper[4945]: I1008 17:05:20.025150 4945 scope.go:117] "RemoveContainer" containerID="30272b8b4955af7eaf4c7c37f239743ee3cf22985d87f28728e95374e8784fa6" Oct 08 17:05:20 crc kubenswrapper[4945]: I1008 17:05:20.472641 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jp64g" event={"ID":"e39fc950-7f86-4d6b-b4b0-25daafb66250","Type":"ContainerStarted","Data":"68083c050f0c5a8559ec6bc277cf3bdf9bf387d607cb51c6a1ec785bfb85b8b6"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071514753024455 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071514754017373 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071477201016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071477201015461 5ustar corecore